Compare commits
1 Commits
master
...
feat-repo+
Author | SHA1 | Date | |
---|---|---|---|
|
7ea8505873 |
3
.gitignore
vendored
3
.gitignore
vendored
@ -33,5 +33,4 @@ lerna-debug.log*
|
|||||||
!.vscode/launch.json
|
!.vscode/launch.json
|
||||||
!.vscode/extensions.json
|
!.vscode/extensions.json
|
||||||
|
|
||||||
/config.yml
|
/config.yml
|
||||||
tsconfig.build.tsbuildinfo
|
|
3
.vscode/extensions.json
vendored
3
.vscode/extensions.json
vendored
@ -1,3 +0,0 @@
|
|||||||
{
|
|
||||||
"recommendations": []
|
|
||||||
}
|
|
0
.vscode/launch.json
vendored
0
.vscode/launch.json
vendored
17
.vscode/settings.json
vendored
17
.vscode/settings.json
vendored
@ -1,20 +1,5 @@
|
|||||||
{
|
{
|
||||||
"cSpell.words": [
|
"cSpell.words": [
|
||||||
"Mutex",
|
"Repos"
|
||||||
"Repos",
|
|
||||||
"amqp",
|
|
||||||
"boardcat",
|
|
||||||
"errout",
|
|
||||||
"fanout",
|
|
||||||
"gitea",
|
|
||||||
"golevelup",
|
|
||||||
"lpush",
|
|
||||||
"lrange",
|
|
||||||
"metatype",
|
|
||||||
"pmessage",
|
|
||||||
"psubscribe",
|
|
||||||
"rabbitmq",
|
|
||||||
"rpop",
|
|
||||||
"rpush"
|
|
||||||
]
|
]
|
||||||
}
|
}
|
@ -1,7 +1,6 @@
|
|||||||
env: dev
|
env: dev
|
||||||
http:
|
http:
|
||||||
port: 7122
|
port: 7122
|
||||||
|
|
||||||
db:
|
db:
|
||||||
postgres:
|
postgres:
|
||||||
host: 192.168.31.194
|
host: 192.168.31.194
|
||||||
@ -9,16 +8,9 @@ db:
|
|||||||
database: fennec
|
database: fennec
|
||||||
username: fennec
|
username: fennec
|
||||||
password:
|
password:
|
||||||
redis:
|
redis:
|
||||||
host: 192.168.31.194
|
mq:
|
||||||
|
host: localhost
|
||||||
port: 6379
|
port: 6379
|
||||||
password:
|
|
||||||
prefix: fennec
|
|
||||||
rabbitmq:
|
|
||||||
uri: 'amqp://fennec:fennec@192.168.31.194:5672'
|
|
||||||
etcd:
|
|
||||||
hosts:
|
|
||||||
- 'http://192.168.31.194:2379'
|
|
||||||
|
|
||||||
workspaces:
|
workspaces:
|
||||||
root: '/Users/ivanli/Projects/fennec/workspaces'
|
root: '/Users/ivanli/Projects/fennec/workspaces'
|
@ -1,33 +0,0 @@
|
|||||||
# CI/CD 流程
|
|
||||||
0. 准备
|
|
||||||
- project information
|
|
||||||
- commit hash
|
|
||||||
1. checkout
|
|
||||||
2. install dependencies
|
|
||||||
3. run test script
|
|
||||||
5. run deploy script
|
|
||||||
6. clear workspace
|
|
||||||
|
|
||||||
## 流水线任务单元描述
|
|
||||||
```json
|
|
||||||
{
|
|
||||||
"version": 1,
|
|
||||||
"unit": {
|
|
||||||
"install-dependencies": {
|
|
||||||
"script": "npm ci"
|
|
||||||
},
|
|
||||||
"test": {
|
|
||||||
"script": "npm test"
|
|
||||||
},
|
|
||||||
"build": {
|
|
||||||
"script": "npm build"
|
|
||||||
},
|
|
||||||
"deploy": {
|
|
||||||
"script": [
|
|
||||||
"npm build"
|
|
||||||
]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
```
|
|
@ -1,14 +0,0 @@
|
|||||||
module.exports = {
|
|
||||||
apps: [
|
|
||||||
{
|
|
||||||
name: 'fennec-be',
|
|
||||||
script: 'npm',
|
|
||||||
args: 'run start:prod',
|
|
||||||
watch: false,
|
|
||||||
ignore_watch: ['node_modules'],
|
|
||||||
log_date_format: 'MM-DD HH:mm:ss.SSS Z',
|
|
||||||
env: {},
|
|
||||||
max_restarts: 5,
|
|
||||||
},
|
|
||||||
],
|
|
||||||
};
|
|
30101
package-lock.json
generated
30101
package-lock.json
generated
File diff suppressed because it is too large
Load Diff
37
package.json
37
package.json
@ -1,8 +1,8 @@
|
|||||||
{
|
{
|
||||||
"name": "fennec-be",
|
"name": "fennec-be",
|
||||||
"version": "0.1.1",
|
"version": "0.0.1",
|
||||||
"description": "a ci/cd project.",
|
"description": "",
|
||||||
"author": "Ivan Li\b<ivanli2048@gmail.com>",
|
"author": "",
|
||||||
"private": true,
|
"private": true,
|
||||||
"license": "UNLICENSED",
|
"license": "UNLICENSED",
|
||||||
"scripts": {
|
"scripts": {
|
||||||
@ -10,8 +10,8 @@
|
|||||||
"build": "nest build",
|
"build": "nest build",
|
||||||
"format": "prettier --write \"src/**/*.ts\" \"test/**/*.ts\"",
|
"format": "prettier --write \"src/**/*.ts\" \"test/**/*.ts\"",
|
||||||
"start": "nest start",
|
"start": "nest start",
|
||||||
"start:dev": "DEBUG=fennec:* nest start --watch",
|
"start:dev": "nest start --watch",
|
||||||
"start:debug": "DEBUG=fennec:* nest start --debug --watch",
|
"start:debug": "DEBUG=simple-git,simple-git:* nest start --debug --watch",
|
||||||
"start:prod": "node dist/main",
|
"start:prod": "node dist/main",
|
||||||
"lint": "eslint \"{src,apps,libs,test}/**/*.ts\" --fix",
|
"lint": "eslint \"{src,apps,libs,test}/**/*.ts\" --fix",
|
||||||
"test": "jest",
|
"test": "jest",
|
||||||
@ -21,34 +21,26 @@
|
|||||||
"test:e2e": "jest --config ./test/jest-e2e.json"
|
"test:e2e": "jest --config ./test/jest-e2e.json"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@golevelup/nestjs-rabbitmq": "^1.16.1",
|
"@nestjs/bull": "^0.3.1",
|
||||||
"@nestjs-lib/auth": "^0.2.1",
|
|
||||||
"@nestjs/common": "^7.5.1",
|
"@nestjs/common": "^7.5.1",
|
||||||
"@nestjs/config": "^0.6.2",
|
"@nestjs/config": "^0.6.2",
|
||||||
"@nestjs/core": "^7.5.1",
|
"@nestjs/core": "^7.5.1",
|
||||||
"@nestjs/graphql": "^7.9.8",
|
"@nestjs/graphql": "^7.9.8",
|
||||||
"@nestjs/platform-express": "^7.5.1",
|
"@nestjs/platform-express": "^7.5.1",
|
||||||
"@nestjs/typeorm": "^7.1.5",
|
"@nestjs/typeorm": "^7.1.5",
|
||||||
"@types/amqplib": "^0.8.0",
|
"@neuralegion/class-sanitizer": "^0.3.2",
|
||||||
|
"@types/bull": "^3.15.0",
|
||||||
"@types/ramda": "^0.27.38",
|
"@types/ramda": "^0.27.38",
|
||||||
"apollo-server-express": "^2.19.2",
|
"apollo-server-express": "^2.19.2",
|
||||||
"bcrypt": "^5.0.0",
|
"bcrypt": "^5.0.0",
|
||||||
"body-parser": "^1.19.0",
|
"bull": "^3.20.1",
|
||||||
"class-transformer": "^0.3.2",
|
"class-transformer": "^0.3.2",
|
||||||
"class-validator": "^0.13.1",
|
"class-validator": "^0.13.1",
|
||||||
"debug": "^4.3.1",
|
|
||||||
"graphql": "^15.5.0",
|
"graphql": "^15.5.0",
|
||||||
"graphql-tools": "^8.1.0",
|
"graphql-tools": "^7.0.2",
|
||||||
"ioredis": "^4.25.0",
|
|
||||||
"jose": "^3.14.0",
|
|
||||||
"js-yaml": "^4.0.0",
|
"js-yaml": "^4.0.0",
|
||||||
"nestjs-etcd": "^0.2.0",
|
|
||||||
"nestjs-pino": "^1.4.0",
|
|
||||||
"nestjs-redis": "^1.2.8",
|
"nestjs-redis": "^1.2.8",
|
||||||
"observable-to-async-generator": "^1.0.1-rc",
|
|
||||||
"pg": "^8.5.1",
|
"pg": "^8.5.1",
|
||||||
"pino-pretty": "^4.7.1",
|
|
||||||
"pm2": "^5.1.0",
|
|
||||||
"ramda": "^0.27.1",
|
"ramda": "^0.27.1",
|
||||||
"reflect-metadata": "^0.1.13",
|
"reflect-metadata": "^0.1.13",
|
||||||
"rimraf": "^3.0.2",
|
"rimraf": "^3.0.2",
|
||||||
@ -57,17 +49,13 @@
|
|||||||
"typeorm": "^0.2.30"
|
"typeorm": "^0.2.30"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@nestjs/cli": "^7.6.0",
|
"@nestjs/cli": "^7.5.1",
|
||||||
"@nestjs/schematics": "^7.1.3",
|
"@nestjs/schematics": "^7.1.3",
|
||||||
"@nestjs/testing": "^7.5.1",
|
"@nestjs/testing": "^7.5.1",
|
||||||
"@types/body-parser": "^1.19.0",
|
|
||||||
"@types/debug": "^4.1.5",
|
|
||||||
"@types/express": "^4.17.8",
|
"@types/express": "^4.17.8",
|
||||||
"@types/ioredis": "^4.22.2",
|
|
||||||
"@types/jest": "^26.0.15",
|
"@types/jest": "^26.0.15",
|
||||||
"@types/js-yaml": "^4.0.0",
|
"@types/js-yaml": "^4.0.0",
|
||||||
"@types/node": "^14.14.6",
|
"@types/node": "^14.14.6",
|
||||||
"@types/pino-pretty": "^4.7.0",
|
|
||||||
"@types/supertest": "^2.0.10",
|
"@types/supertest": "^2.0.10",
|
||||||
"@typescript-eslint/eslint-plugin": "^4.6.1",
|
"@typescript-eslint/eslint-plugin": "^4.6.1",
|
||||||
"@typescript-eslint/parser": "^4.6.1",
|
"@typescript-eslint/parser": "^4.6.1",
|
||||||
@ -97,9 +85,6 @@
|
|||||||
"collectCoverageFrom": [
|
"collectCoverageFrom": [
|
||||||
"**/*.(t|j)s"
|
"**/*.(t|j)s"
|
||||||
],
|
],
|
||||||
"moduleNameMapper": {
|
|
||||||
"^jose/(.*)$": "<rootDir>/../node_modules/jose/dist/node/cjs/$1"
|
|
||||||
},
|
|
||||||
"coverageDirectory": "../coverage",
|
"coverageDirectory": "../coverage",
|
||||||
"testEnvironment": "node"
|
"testEnvironment": "node"
|
||||||
}
|
}
|
||||||
|
@ -1,5 +1,4 @@
|
|||||||
import { CommonsModule } from './commons/commons.module';
|
import { Module } from '@nestjs/common';
|
||||||
import { MiddlewareConsumer, Module, NestModule } from '@nestjs/common';
|
|
||||||
import { ConfigModule, ConfigService } from '@nestjs/config';
|
import { ConfigModule, ConfigService } from '@nestjs/config';
|
||||||
import { GraphQLModule } from '@nestjs/graphql';
|
import { GraphQLModule } from '@nestjs/graphql';
|
||||||
import { TypeOrmModule } from '@nestjs/typeorm';
|
import { TypeOrmModule } from '@nestjs/typeorm';
|
||||||
@ -8,41 +7,14 @@ import { AppResolver } from './app.resolver';
|
|||||||
import { AppService } from './app.service';
|
import { AppService } from './app.service';
|
||||||
import { ProjectsModule } from './projects/projects.module';
|
import { ProjectsModule } from './projects/projects.module';
|
||||||
import { ReposModule } from './repos/repos.module';
|
import { ReposModule } from './repos/repos.module';
|
||||||
import { PipelinesModule } from './pipelines/pipelines.module';
|
|
||||||
import { PipelineTasksModule } from './pipeline-tasks/pipeline-tasks.module';
|
|
||||||
import configuration from './commons/config/configuration';
|
import configuration from './commons/config/configuration';
|
||||||
import { RedisModule } from 'nestjs-redis';
|
import { BullModule } from '@nestjs/bull';
|
||||||
import { WebhooksModule } from './webhooks/webhooks.module';
|
|
||||||
import { RawBodyMiddleware } from './commons/middleware/raw-body.middleware';
|
|
||||||
import { GiteaWebhooksController } from './webhooks/gitea-webhooks.controller';
|
|
||||||
import { ParseBodyMiddleware } from './commons/middleware/parse-body.middleware';
|
|
||||||
import { LoggerModule } from 'nestjs-pino';
|
|
||||||
import { EtcdModule } from 'nestjs-etcd';
|
|
||||||
import pinoPretty from 'pino-pretty';
|
|
||||||
import { fromPairs, map, pipe, toPairs } from 'ramda';
|
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [
|
imports: [
|
||||||
ConfigModule.forRoot({
|
ConfigModule.forRoot({
|
||||||
load: [configuration],
|
load: [configuration],
|
||||||
}),
|
}),
|
||||||
LoggerModule.forRootAsync({
|
|
||||||
imports: [ConfigModule],
|
|
||||||
useFactory: (configService: ConfigService) => {
|
|
||||||
const isDev = configService.get<'dev' | 'prod'>('env') === 'dev';
|
|
||||||
return {
|
|
||||||
pinoHttp: {
|
|
||||||
prettyPrint: isDev
|
|
||||||
? {
|
|
||||||
levelFirst: true,
|
|
||||||
}
|
|
||||||
: false,
|
|
||||||
prettifier: pinoPretty,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
},
|
|
||||||
inject: [ConfigService],
|
|
||||||
}),
|
|
||||||
TypeOrmModule.forRootAsync({
|
TypeOrmModule.forRootAsync({
|
||||||
imports: [ConfigModule],
|
imports: [ConfigModule],
|
||||||
useFactory: (configService: ConfigService) => ({
|
useFactory: (configService: ConfigService) => ({
|
||||||
@ -63,59 +35,23 @@ import { fromPairs, map, pipe, toPairs } from 'ramda';
|
|||||||
debug: configService.get<string>('env') !== 'prod',
|
debug: configService.get<string>('env') !== 'prod',
|
||||||
playground: true,
|
playground: true,
|
||||||
autoSchemaFile: true,
|
autoSchemaFile: true,
|
||||||
installSubscriptionHandlers: true,
|
}),
|
||||||
context: ({ req, connection, ...args }) => {
|
inject: [ConfigService],
|
||||||
return connection ? { req: connection.context } : { req };
|
}),
|
||||||
},
|
BullModule.forRootAsync({
|
||||||
subscriptions: {
|
imports: [ConfigModule],
|
||||||
onConnect: (connectionParams: Record<string, string>) => {
|
useFactory: async (configService: ConfigService) => ({
|
||||||
const connectionParamsWithLowerKeys = pipe(
|
redis: {
|
||||||
toPairs,
|
host: configService.get<string>('redis.mq.host', 'localhost'),
|
||||||
map(([key, value]) => [key.toLowerCase(), value]),
|
port: configService.get<number>('redis.mq.port', 6379),
|
||||||
fromPairs,
|
|
||||||
)(connectionParams);
|
|
||||||
|
|
||||||
return {
|
|
||||||
headers: connectionParamsWithLowerKeys,
|
|
||||||
};
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
inject: [ConfigService],
|
inject: [ConfigService],
|
||||||
}),
|
}),
|
||||||
ProjectsModule,
|
ProjectsModule,
|
||||||
ReposModule,
|
ReposModule,
|
||||||
PipelinesModule,
|
|
||||||
PipelineTasksModule,
|
|
||||||
RedisModule.forRootAsync({
|
|
||||||
imports: [ConfigModule],
|
|
||||||
useFactory: (configService: ConfigService) => ({
|
|
||||||
host: configService.get<string>('db.redis.host', 'localhost'),
|
|
||||||
port: configService.get<number>('db.redis.port', 6379),
|
|
||||||
password: configService.get<string>('db.redis.password', ''),
|
|
||||||
keyPrefix: configService.get<string>('db.redis.prefix', 'fennec') + ':',
|
|
||||||
}),
|
|
||||||
inject: [ConfigService],
|
|
||||||
}),
|
|
||||||
EtcdModule.forRootAsync({
|
|
||||||
imports: [ConfigModule],
|
|
||||||
useFactory: (configService: ConfigService) => ({
|
|
||||||
hosts: configService.get<string>('db.etcd.hosts', 'localhost:2379'),
|
|
||||||
}),
|
|
||||||
inject: [ConfigService],
|
|
||||||
}),
|
|
||||||
WebhooksModule,
|
|
||||||
CommonsModule,
|
|
||||||
],
|
],
|
||||||
controllers: [AppController],
|
controllers: [AppController],
|
||||||
providers: [AppService, AppResolver],
|
providers: [AppService, AppResolver],
|
||||||
})
|
})
|
||||||
export class AppModule implements NestModule {
|
export class AppModule {}
|
||||||
public configure(consumer: MiddlewareConsumer): void {
|
|
||||||
consumer
|
|
||||||
.apply(RawBodyMiddleware)
|
|
||||||
.forRoutes(GiteaWebhooksController)
|
|
||||||
.apply(ParseBodyMiddleware)
|
|
||||||
.forRoutes('*');
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
@ -1,11 +1,8 @@
|
|||||||
import { Module } from '@nestjs/common';
|
import { Module } from '@nestjs/common';
|
||||||
import { PasswordConverter } from './services/password-converter';
|
import { PasswordConverter } from './services/password-converter';
|
||||||
import { RedisMutexModule } from './redis-mutex/redis-mutex.module';
|
|
||||||
import { AuthModule } from '@nestjs-lib/auth';
|
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [RedisMutexModule, AuthModule],
|
|
||||||
providers: [PasswordConverter],
|
providers: [PasswordConverter],
|
||||||
exports: [PasswordConverter, RedisMutexModule, AuthModule],
|
exports: [PasswordConverter],
|
||||||
})
|
})
|
||||||
export class CommonsModule {}
|
export class CommonsModule {}
|
||||||
|
@ -1,7 +1,6 @@
|
|||||||
import { Field, ID, ObjectType } from '@nestjs/graphql';
|
import { Field, ID, ObjectType } from '@nestjs/graphql';
|
||||||
import {
|
import {
|
||||||
CreateDateColumn,
|
CreateDateColumn,
|
||||||
DeleteDateColumn,
|
|
||||||
PrimaryGeneratedColumn,
|
PrimaryGeneratedColumn,
|
||||||
UpdateDateColumn,
|
UpdateDateColumn,
|
||||||
} from 'typeorm';
|
} from 'typeorm';
|
||||||
@ -17,7 +16,4 @@ export class AppBaseEntity {
|
|||||||
|
|
||||||
@UpdateDateColumn({ select: false })
|
@UpdateDateColumn({ select: false })
|
||||||
updatedAt: Date;
|
updatedAt: Date;
|
||||||
|
|
||||||
@DeleteDateColumn({ select: false })
|
|
||||||
deletedAt?: Date;
|
|
||||||
}
|
}
|
||||||
|
@ -1,11 +1,11 @@
|
|||||||
import { pick } from 'ramda';
|
|
||||||
|
|
||||||
export class ApplicationException extends Error {
|
export class ApplicationException extends Error {
|
||||||
code: number;
|
code: number;
|
||||||
error: Error;
|
error: Error;
|
||||||
|
|
||||||
constructor(
|
constructor(
|
||||||
message: string | { error?: Error; message?: string | any; code?: number },
|
message:
|
||||||
|
| string
|
||||||
|
| { error?: Error; message?: string | object; code?: number },
|
||||||
) {
|
) {
|
||||||
if (message instanceof Object) {
|
if (message instanceof Object) {
|
||||||
super();
|
super();
|
||||||
@ -18,8 +18,4 @@ export class ApplicationException extends Error {
|
|||||||
super((message as unknown) as any);
|
super((message as unknown) as any);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
toJSON() {
|
|
||||||
return pick(['code', 'message'], this);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
@ -1,58 +1,55 @@
|
|||||||
import {
|
import {
|
||||||
ExceptionFilter,
|
|
||||||
Catch,
|
|
||||||
ArgumentsHost,
|
ArgumentsHost,
|
||||||
|
Catch,
|
||||||
|
ExceptionFilter,
|
||||||
HttpException,
|
HttpException,
|
||||||
HttpStatus,
|
HttpStatus,
|
||||||
} from '@nestjs/common';
|
} from '@nestjs/common';
|
||||||
import { ApolloError } from 'apollo-server-errors';
|
import { EntityNotFoundError } from 'typeorm/error/EntityNotFoundError';
|
||||||
import { PinoLogger, InjectPinoLogger } from 'nestjs-pino';
|
|
||||||
|
|
||||||
@Catch(HttpException)
|
@Catch()
|
||||||
export class HttpExceptionFilter implements ExceptionFilter {
|
export class AllExceptionsFilter implements ExceptionFilter {
|
||||||
constructor(
|
catch(exception: any, host: ArgumentsHost) {
|
||||||
@InjectPinoLogger(HttpExceptionFilter.name)
|
const ctx = host.switchToHttp();
|
||||||
private readonly logger: PinoLogger,
|
const response = ctx.getResponse();
|
||||||
) {}
|
const request = ctx.getRequest();
|
||||||
catch(exception: HttpException, host: ArgumentsHost) {
|
|
||||||
switch (host.getType<'http' | 'graphql' | string>()) {
|
|
||||||
case 'graphql': {
|
|
||||||
const errorName = exception.message;
|
|
||||||
const extensions: Record<string, any> = {};
|
|
||||||
const err = exception.getResponse();
|
|
||||||
if (typeof err === 'string') {
|
|
||||||
extensions.message = err;
|
|
||||||
} else {
|
|
||||||
Object.assign(extensions, (err as any).extension);
|
|
||||||
extensions.message = (err as any).message;
|
|
||||||
}
|
|
||||||
extensions.error = errorName;
|
|
||||||
this.logger.error(extensions);
|
|
||||||
return new ApolloError(
|
|
||||||
extensions.message,
|
|
||||||
exception.getStatus().toString(),
|
|
||||||
extensions,
|
|
||||||
);
|
|
||||||
}
|
|
||||||
case 'http': {
|
|
||||||
const ctx = host.switchToHttp();
|
|
||||||
const response = ctx.getResponse();
|
|
||||||
const request = ctx.getRequest();
|
|
||||||
|
|
||||||
const status =
|
const status =
|
||||||
exception instanceof HttpException
|
exception instanceof HttpException
|
||||||
? exception.getStatus()
|
? exception.getStatus()
|
||||||
: HttpStatus.INTERNAL_SERVER_ERROR;
|
: HttpStatus.INTERNAL_SERVER_ERROR;
|
||||||
|
|
||||||
|
if (exception instanceof HttpException) {
|
||||||
|
const ex = exception.getResponse();
|
||||||
|
if (ex instanceof Object) {
|
||||||
response.status(status).json({
|
response.status(status).json({
|
||||||
statusCode: status,
|
...ex,
|
||||||
message: exception.message,
|
timestamp: Date.now(),
|
||||||
timestamp: new Date().toISOString(),
|
path: request.url,
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
response.status(status).json({
|
||||||
|
message: ex,
|
||||||
|
timestamp: Date.now(),
|
||||||
path: request.url,
|
path: request.url,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
default:
|
} else if (exception instanceof EntityNotFoundError) {
|
||||||
throw exception;
|
response.status(HttpStatus.NOT_FOUND).json({
|
||||||
|
message: '资源未找到!',
|
||||||
|
timestamp: Date.now(),
|
||||||
|
path: request.url,
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
console.error('服务器内部错误');
|
||||||
|
console.error(exception);
|
||||||
|
response.status(HttpStatus.INTERNAL_SERVER_ERROR).json({
|
||||||
|
code: status,
|
||||||
|
timestamp: new Date().toISOString(),
|
||||||
|
message: '服务器内部错误',
|
||||||
|
error: exception,
|
||||||
|
path: request.url,
|
||||||
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,7 +0,0 @@
|
|||||||
import { ParseBodyMiddleware } from './parse-body.middleware';
|
|
||||||
|
|
||||||
describe('ParseBodyMiddleware', () => {
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(new ParseBodyMiddleware()).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,13 +0,0 @@
|
|||||||
import { Injectable, NestMiddleware } from '@nestjs/common';
|
|
||||||
import { json, urlencoded, text } from 'body-parser';
|
|
||||||
import { Request, Response, NextFunction } from 'express';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class ParseBodyMiddleware implements NestMiddleware {
|
|
||||||
use(req: Request, res: Response, next: NextFunction) {
|
|
||||||
json()(req, res, () =>
|
|
||||||
urlencoded()(req, res, () => text()(req, res, next)),
|
|
||||||
);
|
|
||||||
// next();
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,7 +0,0 @@
|
|||||||
import { RawBodyMiddleware } from './raw-body.middleware';
|
|
||||||
|
|
||||||
describe('RawBodyMiddleware', () => {
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(new RawBodyMiddleware()).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,10 +0,0 @@
|
|||||||
import { Injectable, NestMiddleware } from '@nestjs/common';
|
|
||||||
import { raw } from 'body-parser';
|
|
||||||
import { Request, Response, NextFunction } from 'express';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class RawBodyMiddleware implements NestMiddleware {
|
|
||||||
use(req: Request, res: Response, next: NextFunction) {
|
|
||||||
raw({ type: '*/*' })(req, res, next);
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,25 +0,0 @@
|
|||||||
import { ArgumentMetadata, Injectable, PipeTransform } from '@nestjs/common';
|
|
||||||
import { plainToClass } from 'class-transformer';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class SanitizePipe implements PipeTransform {
|
|
||||||
transform(value: any, metadata: ArgumentMetadata) {
|
|
||||||
if (
|
|
||||||
!(value instanceof Object) ||
|
|
||||||
value instanceof Buffer ||
|
|
||||||
value instanceof Array
|
|
||||||
) {
|
|
||||||
return value;
|
|
||||||
}
|
|
||||||
const constructorFunction = metadata.metatype;
|
|
||||||
if (!constructorFunction || value instanceof constructorFunction) {
|
|
||||||
return value;
|
|
||||||
}
|
|
||||||
try {
|
|
||||||
return plainToClass(constructorFunction, value);
|
|
||||||
} catch (err) {
|
|
||||||
console.error(err);
|
|
||||||
throw err;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,10 +0,0 @@
|
|||||||
import { Module } from '@nestjs/common';
|
|
||||||
import { RedisMutexService } from './redis-mutex.service';
|
|
||||||
import { RedisModule } from 'nestjs-redis';
|
|
||||||
|
|
||||||
@Module({
|
|
||||||
imports: [RedisModule],
|
|
||||||
providers: [RedisMutexService],
|
|
||||||
exports: [RedisMutexService],
|
|
||||||
})
|
|
||||||
export class RedisMutexModule {}
|
|
@ -1,25 +0,0 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { RedisService } from 'nestjs-redis';
|
|
||||||
import { RedisMutexService } from './redis-mutex.service';
|
|
||||||
|
|
||||||
describe('RedisMutexService', () => {
|
|
||||||
let service: RedisMutexService;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
RedisMutexService,
|
|
||||||
{
|
|
||||||
provide: RedisService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
service = module.get<RedisMutexService>(RedisMutexService);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(service).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,71 +0,0 @@
|
|||||||
import { Injectable } from '@nestjs/common';
|
|
||||||
import { RedisService } from 'nestjs-redis';
|
|
||||||
import * as uuid from 'uuid';
|
|
||||||
import { ApplicationException } from '../exceptions/application.exception';
|
|
||||||
|
|
||||||
export interface RedisMutexOption {
|
|
||||||
/**
|
|
||||||
* seconds
|
|
||||||
*/
|
|
||||||
expires?: number;
|
|
||||||
/**
|
|
||||||
* seconds
|
|
||||||
*/
|
|
||||||
timeout?: number | null;
|
|
||||||
/**
|
|
||||||
* milliseconds
|
|
||||||
*/
|
|
||||||
retryDelay?: number;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class RedisMutexService {
|
|
||||||
constructor(private readonly redisClient: RedisService) {}
|
|
||||||
|
|
||||||
public async lock(
|
|
||||||
key: string,
|
|
||||||
{ expires = 100, timeout = 10, retryDelay = 100 }: RedisMutexOption = {
|
|
||||||
expires: 100,
|
|
||||||
timeout: 10,
|
|
||||||
retryDelay: 100,
|
|
||||||
},
|
|
||||||
) {
|
|
||||||
const redisKey = `${'mutex-lock'}:${key}`;
|
|
||||||
const redis = this.redisClient.getClient();
|
|
||||||
const value = uuid.v4();
|
|
||||||
const timeoutAt = timeout ? Date.now() + timeout * 1000 : null;
|
|
||||||
|
|
||||||
while (
|
|
||||||
!(await redis
|
|
||||||
.set(redisKey, value, 'EX', expires, 'NX')
|
|
||||||
.then(() => true)
|
|
||||||
.catch(() => false))
|
|
||||||
) {
|
|
||||||
if (timeoutAt && timeoutAt > Date.now()) {
|
|
||||||
throw new ApplicationException('lock timeout');
|
|
||||||
}
|
|
||||||
await new Promise((resolve) => setTimeout(resolve, retryDelay));
|
|
||||||
}
|
|
||||||
|
|
||||||
const renewTimer = setInterval(() => {
|
|
||||||
redis.expire(redisKey, expires);
|
|
||||||
}, (expires * 1000) / 2);
|
|
||||||
|
|
||||||
return async () => {
|
|
||||||
clearInterval(renewTimer);
|
|
||||||
await redis.eval(
|
|
||||||
`
|
|
||||||
if redis.call("get", KEYS[1]) == ARGV[1]
|
|
||||||
then
|
|
||||||
return redis.call("del", KEYS[1])
|
|
||||||
else
|
|
||||||
return 0
|
|
||||||
end
|
|
||||||
`,
|
|
||||||
1,
|
|
||||||
redisKey,
|
|
||||||
value,
|
|
||||||
);
|
|
||||||
};
|
|
||||||
}
|
|
||||||
}
|
|
@ -53,22 +53,8 @@ export class BaseDbService<Entity extends AppBaseEntity> extends TypeormHelper {
|
|||||||
async isDuplicateEntityForUpdate<Dto extends Entity>(
|
async isDuplicateEntityForUpdate<Dto extends Entity>(
|
||||||
id: string,
|
id: string,
|
||||||
dto: Partial<Dto>,
|
dto: Partial<Dto>,
|
||||||
extendsFields?: Array<keyof Dto & string>,
|
|
||||||
): Promise<false | never>;
|
|
||||||
async isDuplicateEntityForUpdate<Dto extends Entity>(
|
|
||||||
old: Entity,
|
|
||||||
dto: Partial<Dto>,
|
|
||||||
extendsFields?: Array<keyof Dto & string>,
|
|
||||||
): Promise<false | never>;
|
|
||||||
async isDuplicateEntityForUpdate<Dto extends Entity>(
|
|
||||||
id: string | Entity,
|
|
||||||
dto: Partial<Dto>,
|
|
||||||
extendsFields: Array<keyof Dto & string> = [],
|
extendsFields: Array<keyof Dto & string> = [],
|
||||||
): Promise<false | never> {
|
): Promise<false | never> {
|
||||||
if (typeof id !== 'string') {
|
|
||||||
dto = Object.assign({}, id, dto);
|
|
||||||
id = id.id;
|
|
||||||
}
|
|
||||||
const qb = this.repository.createQueryBuilder('entity');
|
const qb = this.repository.createQueryBuilder('entity');
|
||||||
const compareFields = this.getCompareFields(dto, [
|
const compareFields = this.getCompareFields(dto, [
|
||||||
...this.uniqueFields,
|
...this.uniqueFields,
|
||||||
|
@ -1,21 +0,0 @@
|
|||||||
import { hostname } from 'os';
|
|
||||||
|
|
||||||
export function getInstanceName() {
|
|
||||||
return hostname();
|
|
||||||
}
|
|
||||||
|
|
||||||
export function getSelfInstanceRouteKey(key: string) {
|
|
||||||
return getAppInstanceRouteKey(key, getInstanceName());
|
|
||||||
}
|
|
||||||
|
|
||||||
export function getAppInstanceRouteKey(key: string, appInstance?: string) {
|
|
||||||
return appInstance ? `${key}.${appInstance}` : key;
|
|
||||||
}
|
|
||||||
|
|
||||||
export function getSelfInstanceQueueKey(key: string) {
|
|
||||||
return getAppInstanceQueueKey(key, getInstanceName());
|
|
||||||
}
|
|
||||||
|
|
||||||
export function getAppInstanceQueueKey(key: string, appInstance?: string) {
|
|
||||||
return appInstance ? `${key}.${appInstance}` : key;
|
|
||||||
}
|
|
14
src/main.ts
14
src/main.ts
@ -1,22 +1,12 @@
|
|||||||
import { PinoLogger } from 'nestjs-pino';
|
|
||||||
import { ValidationPipe } from '@nestjs/common';
|
import { ValidationPipe } from '@nestjs/common';
|
||||||
import { ConfigService } from '@nestjs/config';
|
import { ConfigService } from '@nestjs/config';
|
||||||
import { NestFactory } from '@nestjs/core';
|
import { NestFactory } from '@nestjs/core';
|
||||||
import { AppModule } from './app.module';
|
import { AppModule } from './app.module';
|
||||||
import { HttpExceptionFilter } from './commons/filters/all.exception-filter';
|
|
||||||
import { SanitizePipe } from './commons/pipes/sanitize.pipe';
|
|
||||||
|
|
||||||
async function bootstrap() {
|
async function bootstrap() {
|
||||||
const app = await NestFactory.create(AppModule, { bodyParser: false });
|
const app = await NestFactory.create(AppModule);
|
||||||
const configService = app.get(ConfigService);
|
const configService = app.get(ConfigService);
|
||||||
app.useGlobalPipes(new SanitizePipe());
|
app.useGlobalPipes(new ValidationPipe());
|
||||||
app.useGlobalPipes(
|
|
||||||
new ValidationPipe({
|
|
||||||
transform: true,
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
const httpExceptionFilterLogger = await app.resolve(PinoLogger);
|
|
||||||
app.useGlobalFilters(new HttpExceptionFilter(httpExceptionFilterLogger));
|
|
||||||
await app.listen(configService.get<number>('http.port'));
|
await app.listen(configService.get<number>('http.port'));
|
||||||
}
|
}
|
||||||
bootstrap();
|
bootstrap();
|
||||||
|
@ -1,11 +0,0 @@
|
|||||||
import { Field, InputType } from '@nestjs/graphql';
|
|
||||||
import { PipelineUnits } from '../enums/pipeline-units.enum';
|
|
||||||
|
|
||||||
@InputType()
|
|
||||||
export class CreatePipelineTaskInput {
|
|
||||||
pipelineId: string;
|
|
||||||
|
|
||||||
commit: string;
|
|
||||||
|
|
||||||
units: PipelineUnits[];
|
|
||||||
}
|
|
@ -1,8 +0,0 @@
|
|||||||
import { ArgsType } from '@nestjs/graphql';
|
|
||||||
import { IsUUID } from 'class-validator';
|
|
||||||
|
|
||||||
@ArgsType()
|
|
||||||
export class PipelineTaskLogArgs {
|
|
||||||
@IsUUID()
|
|
||||||
taskId: string;
|
|
||||||
}
|
|
@ -1,14 +0,0 @@
|
|||||||
import { registerEnumType } from '@nestjs/graphql';
|
|
||||||
|
|
||||||
export enum PipelineUnits {
|
|
||||||
checkout = 'checkout',
|
|
||||||
installDependencies = 'installDependencies',
|
|
||||||
test = 'test',
|
|
||||||
deploy = 'deploy',
|
|
||||||
cleanUp = 'cleanUp',
|
|
||||||
}
|
|
||||||
|
|
||||||
registerEnumType(PipelineUnits, {
|
|
||||||
name: 'PipelineUnits',
|
|
||||||
description: '流水线单元',
|
|
||||||
});
|
|
@ -1,15 +0,0 @@
|
|||||||
import { registerEnumType } from '@nestjs/graphql';
|
|
||||||
|
|
||||||
export enum TaskStatuses {
|
|
||||||
success = 'success',
|
|
||||||
failed = 'failed',
|
|
||||||
working = 'working',
|
|
||||||
pending = 'pending',
|
|
||||||
}
|
|
||||||
|
|
||||||
registerEnumType(TaskStatuses, {
|
|
||||||
name: 'TaskStatuses',
|
|
||||||
description: '任务状态',
|
|
||||||
});
|
|
||||||
|
|
||||||
export const terminalTaskStatuses = [TaskStatuses.success, TaskStatuses.failed];
|
|
@ -1,25 +0,0 @@
|
|||||||
import { Field, ObjectType } from '@nestjs/graphql';
|
|
||||||
import { PipelineUnits } from '../enums/pipeline-units.enum';
|
|
||||||
import { TaskStatuses } from '../enums/task-statuses.enum';
|
|
||||||
import { Type } from 'class-transformer';
|
|
||||||
|
|
||||||
@ObjectType()
|
|
||||||
export class PipelineTaskEvent {
|
|
||||||
@Field()
|
|
||||||
taskId: string;
|
|
||||||
@Field()
|
|
||||||
pipelineId: string;
|
|
||||||
@Field()
|
|
||||||
projectId: string;
|
|
||||||
@Field(() => PipelineUnits, { nullable: true })
|
|
||||||
unit: PipelineUnits | null;
|
|
||||||
@Field()
|
|
||||||
@Type(() => Date)
|
|
||||||
emittedAt: Date;
|
|
||||||
@Field()
|
|
||||||
message: string;
|
|
||||||
@Field()
|
|
||||||
messageType: 'stdout' | 'stderr' | 'stdin';
|
|
||||||
@Field(() => TaskStatuses)
|
|
||||||
status: TaskStatuses;
|
|
||||||
}
|
|
@ -1,34 +0,0 @@
|
|||||||
import { PipelineTask } from './../pipeline-task.entity';
|
|
||||||
import { PipelineUnits } from '../enums/pipeline-units.enum';
|
|
||||||
import { Field, HideField, ObjectType } from '@nestjs/graphql';
|
|
||||||
import { Type } from 'class-transformer';
|
|
||||||
|
|
||||||
@ObjectType()
|
|
||||||
export class PipelineTaskLogMessage {
|
|
||||||
@HideField()
|
|
||||||
task: PipelineTask;
|
|
||||||
@Field(() => PipelineUnits, { nullable: true })
|
|
||||||
unit?: PipelineUnits;
|
|
||||||
@Field()
|
|
||||||
@Type(() => Date)
|
|
||||||
time: Date;
|
|
||||||
@Field()
|
|
||||||
message: string;
|
|
||||||
@Field()
|
|
||||||
isError: boolean;
|
|
||||||
|
|
||||||
static create(
|
|
||||||
task: PipelineTask,
|
|
||||||
unit: PipelineUnits,
|
|
||||||
message: string,
|
|
||||||
isError: boolean,
|
|
||||||
) {
|
|
||||||
return Object.assign(new PipelineTaskLogMessage(), {
|
|
||||||
task,
|
|
||||||
message,
|
|
||||||
time: new Date(),
|
|
||||||
unit,
|
|
||||||
isError,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,17 +0,0 @@
|
|||||||
import { TaskStatuses } from '../enums/task-statuses.enum';
|
|
||||||
import { PipelineUnits } from '../enums/pipeline-units.enum';
|
|
||||||
import { Field, ObjectType } from '@nestjs/graphql';
|
|
||||||
import { Type } from 'class-transformer';
|
|
||||||
|
|
||||||
@ObjectType()
|
|
||||||
export class PipelineTaskLogs {
|
|
||||||
@Field(() => PipelineUnits)
|
|
||||||
unit: PipelineUnits;
|
|
||||||
@Field(() => TaskStatuses)
|
|
||||||
status: TaskStatuses;
|
|
||||||
@Type(() => Date)
|
|
||||||
startedAt?: Date;
|
|
||||||
@Type(() => Date)
|
|
||||||
endedAt?: Date;
|
|
||||||
logs = '';
|
|
||||||
}
|
|
@ -1,16 +0,0 @@
|
|||||||
import { Field, InputType, Int, ObjectType } from '@nestjs/graphql';
|
|
||||||
import { Type } from 'class-transformer';
|
|
||||||
import { IsInstance, isInstance, ValidateNested } from 'class-validator';
|
|
||||||
import { WorkUnit } from './work-unit.model';
|
|
||||||
|
|
||||||
@InputType('WorkUnitMetadataInput')
|
|
||||||
@ObjectType()
|
|
||||||
export class WorkUnitMetadata {
|
|
||||||
@Field(() => Int)
|
|
||||||
version = 1;
|
|
||||||
|
|
||||||
@Type(() => WorkUnit)
|
|
||||||
@IsInstance(WorkUnit, { each: true })
|
|
||||||
@ValidateNested({ each: true })
|
|
||||||
units: WorkUnit[];
|
|
||||||
}
|
|
@ -1,16 +0,0 @@
|
|||||||
import { Field, InputType, ObjectType } from '@nestjs/graphql';
|
|
||||||
import { IsNotEmpty } from 'class-validator';
|
|
||||||
import {
|
|
||||||
PipelineUnits,
|
|
||||||
PipelineUnits as PipelineUnitTypes,
|
|
||||||
} from '../enums/pipeline-units.enum';
|
|
||||||
|
|
||||||
@ObjectType()
|
|
||||||
@InputType('WorkUnitInput')
|
|
||||||
export class WorkUnit {
|
|
||||||
@Field(() => PipelineUnits)
|
|
||||||
type: PipelineUnitTypes;
|
|
||||||
|
|
||||||
@IsNotEmpty({ each: true })
|
|
||||||
scripts: string[];
|
|
||||||
}
|
|
@ -1,88 +0,0 @@
|
|||||||
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { RedisService } from 'nestjs-redis';
|
|
||||||
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
|
|
||||||
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
|
||||||
import { TaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
ROUTE_PIPELINE_TASK_DONE,
|
|
||||||
} from './pipeline-tasks.constants';
|
|
||||||
|
|
||||||
describe('PipelineTaskFlushService', () => {
|
|
||||||
let service: PipelineTaskFlushService;
|
|
||||||
let redisService: RedisService;
|
|
||||||
let amqpConnection: AmqpConnection;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const redisClient = {
|
|
||||||
rpush: jest.fn(() => Promise.resolve()),
|
|
||||||
lrange: jest.fn(() => Promise.resolve()),
|
|
||||||
expire: jest.fn(() => Promise.resolve()),
|
|
||||||
};
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
PipelineTaskFlushService,
|
|
||||||
{
|
|
||||||
provide: RedisService,
|
|
||||||
useValue: {
|
|
||||||
getClient() {
|
|
||||||
return redisClient;
|
|
||||||
},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: AmqpConnection,
|
|
||||||
useValue: {
|
|
||||||
request: jest.fn(() => Promise.resolve()),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
service = module.get<PipelineTaskFlushService>(PipelineTaskFlushService);
|
|
||||||
redisService = module.get<RedisService>(RedisService);
|
|
||||||
amqpConnection = module.get<AmqpConnection>(AmqpConnection);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(service).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('write', () => {
|
|
||||||
const amqpMsg = {
|
|
||||||
properties: { headers: { sender: 'test' } },
|
|
||||||
} as any;
|
|
||||||
it('normal', async () => {
|
|
||||||
const testEvent = new PipelineTaskEvent();
|
|
||||||
testEvent.taskId = 'test';
|
|
||||||
testEvent.status = TaskStatuses.working;
|
|
||||||
const rpush = jest.spyOn(redisService.getClient(), 'rpush');
|
|
||||||
const request = jest.spyOn(amqpConnection, 'request');
|
|
||||||
await service.write(testEvent, amqpMsg);
|
|
||||||
expect(rpush).toBeCalledTimes(1);
|
|
||||||
expect(rpush.mock.calls[0][0]).toEqual('p-task:log:test');
|
|
||||||
expect(rpush.mock.calls[0][1]).toEqual(JSON.stringify(testEvent));
|
|
||||||
expect(request).toBeCalledTimes(1);
|
|
||||||
});
|
|
||||||
it('event for which task done', async () => {
|
|
||||||
const testEvent = new PipelineTaskEvent();
|
|
||||||
testEvent.taskId = 'test';
|
|
||||||
testEvent.status = TaskStatuses.success;
|
|
||||||
const rpush = jest.spyOn(redisService.getClient(), 'rpush');
|
|
||||||
const request = jest.spyOn(amqpConnection, 'request');
|
|
||||||
await service.write(testEvent, amqpMsg);
|
|
||||||
expect(rpush).toBeCalledTimes(1);
|
|
||||||
expect(request).toBeCalledTimes(1);
|
|
||||||
expect(request.mock.calls[0][0]).toMatchObject({
|
|
||||||
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
routingKey: ROUTE_PIPELINE_TASK_DONE,
|
|
||||||
payload: {
|
|
||||||
taskId: 'test',
|
|
||||||
status: TaskStatuses.success,
|
|
||||||
runOn: 'test',
|
|
||||||
},
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,66 +0,0 @@
|
|||||||
import { AmqpConnection, RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { Injectable } from '@nestjs/common';
|
|
||||||
import { ConsumeMessage } from 'amqplib';
|
|
||||||
import { deserialize } from 'class-transformer';
|
|
||||||
import { RedisService } from 'nestjs-redis';
|
|
||||||
import { isNil } from 'ramda';
|
|
||||||
import { getSelfInstanceQueueKey } from '../commons/utils/rabbit-mq';
|
|
||||||
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
ROUTE_PIPELINE_TASK_DONE,
|
|
||||||
} from './pipeline-tasks.constants';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PIPELINE_TASK_FANOUT,
|
|
||||||
ROUTE_PIPELINE_TASK_LOG,
|
|
||||||
QUEUE_WRITE_PIPELINE_TASK_LOG,
|
|
||||||
} from './pipeline-tasks.constants';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class PipelineTaskFlushService {
|
|
||||||
constructor(
|
|
||||||
private readonly redisService: RedisService,
|
|
||||||
private readonly amqpConnection: AmqpConnection,
|
|
||||||
) {}
|
|
||||||
|
|
||||||
@RabbitSubscribe({
|
|
||||||
exchange: EXCHANGE_PIPELINE_TASK_FANOUT,
|
|
||||||
routingKey: ROUTE_PIPELINE_TASK_LOG,
|
|
||||||
queue: getSelfInstanceQueueKey(QUEUE_WRITE_PIPELINE_TASK_LOG),
|
|
||||||
queueOptions: {
|
|
||||||
autoDelete: true,
|
|
||||||
durable: true,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async write(message: PipelineTaskEvent, amqpMsg: ConsumeMessage) {
|
|
||||||
const client = this.redisService.getClient();
|
|
||||||
await client.rpush(this.getKey(message.taskId), JSON.stringify(message));
|
|
||||||
await client.expire(this.getKey(message.taskId), 600); // ten minutes
|
|
||||||
if (isNil(message.unit)) {
|
|
||||||
try {
|
|
||||||
await this.amqpConnection.request({
|
|
||||||
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
routingKey: ROUTE_PIPELINE_TASK_DONE,
|
|
||||||
payload: {
|
|
||||||
taskId: message.taskId,
|
|
||||||
status: message.status,
|
|
||||||
runOn: amqpMsg.properties.headers.sender,
|
|
||||||
},
|
|
||||||
});
|
|
||||||
} catch (error) {
|
|
||||||
console.log(error);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async read(taskId: string) {
|
|
||||||
const raw = await this.redisService
|
|
||||||
.getClient()
|
|
||||||
.lrange(this.getKey(taskId), 0, -1);
|
|
||||||
return raw.map((it) => deserialize(PipelineTaskEvent, it));
|
|
||||||
}
|
|
||||||
|
|
||||||
private getKey(taskId: string) {
|
|
||||||
return `p-task:log:${taskId}`;
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,42 +0,0 @@
|
|||||||
import { AppBaseEntity } from './../commons/entities/app-base-entity';
|
|
||||||
import { ObjectType } from '@nestjs/graphql';
|
|
||||||
import { Column, Entity, ManyToOne, ValueTransformer } from 'typeorm';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
|
||||||
import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
|
|
||||||
import { TaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import { PipelineUnits } from './enums/pipeline-units.enum';
|
|
||||||
import { plainToClass } from 'class-transformer';
|
|
||||||
|
|
||||||
const logsTransformer: ValueTransformer = {
|
|
||||||
from: (value) => plainToClass(PipelineTaskLogs, value),
|
|
||||||
to: (value) => value,
|
|
||||||
};
|
|
||||||
@ObjectType()
|
|
||||||
@Entity()
|
|
||||||
export class PipelineTask extends AppBaseEntity {
|
|
||||||
@ManyToOne(() => Pipeline)
|
|
||||||
pipeline: Pipeline;
|
|
||||||
@Column()
|
|
||||||
pipelineId: string;
|
|
||||||
|
|
||||||
@Column()
|
|
||||||
commit: string;
|
|
||||||
|
|
||||||
@Column({ type: 'enum', enum: PipelineUnits, array: true })
|
|
||||||
units: PipelineUnits[];
|
|
||||||
|
|
||||||
@Column({ type: 'jsonb', default: '[]', transformer: logsTransformer })
|
|
||||||
logs: PipelineTaskLogs[];
|
|
||||||
|
|
||||||
@Column({ type: 'enum', enum: TaskStatuses, default: TaskStatuses.pending })
|
|
||||||
status: TaskStatuses;
|
|
||||||
|
|
||||||
@Column({ nullable: true })
|
|
||||||
startedAt?: Date;
|
|
||||||
|
|
||||||
@Column({ nullable: true })
|
|
||||||
endedAt?: Date;
|
|
||||||
|
|
||||||
@Column({ nullable: true })
|
|
||||||
runOn: string;
|
|
||||||
}
|
|
@ -1,77 +0,0 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { PipelineTaskLogger } from './pipeline-task.logger';
|
|
||||||
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
|
||||||
import { take, timeout } from 'rxjs/operators';
|
|
||||||
|
|
||||||
describe('PipelineTaskRunner', () => {
|
|
||||||
let logger: PipelineTaskLogger;
|
|
||||||
let module: TestingModule;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
module = await Test.createTestingModule({
|
|
||||||
providers: [PipelineTaskLogger],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
logger = module.get(PipelineTaskLogger);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(logger).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('getMessage$', () => {
|
|
||||||
it('normal', async () => {
|
|
||||||
const event = new PipelineTaskEvent();
|
|
||||||
event.taskId = 'test';
|
|
||||||
const emittedAt = new Date();
|
|
||||||
event.emittedAt = emittedAt.toISOString() as any;
|
|
||||||
const message$ = logger.getMessage$('test');
|
|
||||||
|
|
||||||
let receiveEvent;
|
|
||||||
message$.pipe(take(1)).subscribe((value) => (receiveEvent = value));
|
|
||||||
await logger.handleEvent(event);
|
|
||||||
expect(receiveEvent).toMatchObject({
|
|
||||||
...event,
|
|
||||||
emittedAt,
|
|
||||||
});
|
|
||||||
});
|
|
||||||
it('no match', async () => {
|
|
||||||
const event = new PipelineTaskEvent();
|
|
||||||
event.taskId = 'test';
|
|
||||||
const message$ = logger.getMessage$('other');
|
|
||||||
setTimeout(() => {
|
|
||||||
logger.handleEvent(event);
|
|
||||||
});
|
|
||||||
await expect(
|
|
||||||
message$.pipe(take(1), timeout(100)).toPromise(),
|
|
||||||
).rejects.toThrow(/timeout/i);
|
|
||||||
});
|
|
||||||
it('multiple subscribers', async () => {
|
|
||||||
const event = new PipelineTaskEvent();
|
|
||||||
event.taskId = 'test';
|
|
||||||
const message$ = logger.getMessage$('test');
|
|
||||||
const message2$ = logger.getMessage$('test');
|
|
||||||
setTimeout(() => {
|
|
||||||
logger.handleEvent(event);
|
|
||||||
logger.handleEvent(event);
|
|
||||||
});
|
|
||||||
await Promise.all([
|
|
||||||
expect(
|
|
||||||
message$.pipe(take(1), timeout(100)).toPromise(),
|
|
||||||
).resolves.toEqual(event),
|
|
||||||
expect(
|
|
||||||
message2$.pipe(take(1), timeout(100)).toPromise(),
|
|
||||||
).resolves.toEqual(event),
|
|
||||||
]);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('onModuleDestroy', () => {
|
|
||||||
it('complete observable when destroying module', async () => {
|
|
||||||
logger.onModuleDestroy();
|
|
||||||
await expect(
|
|
||||||
(logger as any).message$.toPromise(),
|
|
||||||
).resolves.toBeUndefined();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,38 +0,0 @@
|
|||||||
import { RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { Injectable, OnModuleDestroy } from '@nestjs/common';
|
|
||||||
import { plainToClass } from 'class-transformer';
|
|
||||||
import { Observable, Subject } from 'rxjs';
|
|
||||||
import { filter } from 'rxjs/operators';
|
|
||||||
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PIPELINE_TASK_FANOUT,
|
|
||||||
QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT,
|
|
||||||
ROUTE_PIPELINE_TASK_LOG,
|
|
||||||
} from './pipeline-tasks.constants';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class PipelineTaskLogger implements OnModuleDestroy {
|
|
||||||
private readonly messageSubject = new Subject<PipelineTaskEvent>();
|
|
||||||
private readonly message$: Observable<PipelineTaskEvent> =
|
|
||||||
this.messageSubject.pipe();
|
|
||||||
|
|
||||||
@RabbitSubscribe({
|
|
||||||
exchange: EXCHANGE_PIPELINE_TASK_FANOUT,
|
|
||||||
routingKey: ROUTE_PIPELINE_TASK_LOG,
|
|
||||||
queue: QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT,
|
|
||||||
queueOptions: {
|
|
||||||
autoDelete: true,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async handleEvent(message: PipelineTaskEvent) {
|
|
||||||
this.messageSubject.next(plainToClass(PipelineTaskEvent, message));
|
|
||||||
}
|
|
||||||
|
|
||||||
getMessage$(taskId: string) {
|
|
||||||
return this.message$.pipe(filter((event) => event.taskId === taskId));
|
|
||||||
}
|
|
||||||
|
|
||||||
onModuleDestroy() {
|
|
||||||
this.messageSubject.complete();
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,351 +0,0 @@
|
|||||||
import { DeployByPm2Service } from './runners/deploy-by-pm2/deploy-by-pm2.service';
|
|
||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { ReposService } from '../repos/repos.service';
|
|
||||||
import { PipelineUnits } from './enums/pipeline-units.enum';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
|
||||||
import { Project } from '../projects/project.entity';
|
|
||||||
import { TaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
|
|
||||||
import { PipelineTaskRunner } from './pipeline-task.runner';
|
|
||||||
import { WorkUnitMetadata } from './models/work-unit-metadata.model';
|
|
||||||
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
describe('PipelineTaskRunner', () => {
|
|
||||||
let runner: PipelineTaskRunner;
|
|
||||||
let deployByPM2Service: DeployByPm2Service;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
{
|
|
||||||
provide: ReposService,
|
|
||||||
useValue: {
|
|
||||||
getWorkspaceRootByTask: () => 'workspace-root',
|
|
||||||
checkout: async () => undefined,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: getLoggerToken(PipelineTaskRunner.name),
|
|
||||||
useValue: new PinoLogger({}),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: 'spawn',
|
|
||||||
useValue: () => undefined,
|
|
||||||
},
|
|
||||||
PipelineTaskRunner,
|
|
||||||
{
|
|
||||||
provide: AmqpConnection,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: DeployByPm2Service,
|
|
||||||
useValue: {
|
|
||||||
deploy: () => Promise.resolve(),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
module.get(ReposService);
|
|
||||||
runner = module.get(PipelineTaskRunner);
|
|
||||||
deployByPM2Service = module.get(DeployByPm2Service);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(runner).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('onNewTask', async () => {
|
|
||||||
const task = new PipelineTask();
|
|
||||||
let tmpTask;
|
|
||||||
const doTask = jest
|
|
||||||
.spyOn(runner, 'doTask')
|
|
||||||
.mockImplementation(async (task) => {
|
|
||||||
tmpTask = task;
|
|
||||||
});
|
|
||||||
await runner.onNewTask(task);
|
|
||||||
expect(tmpTask).toEqual(task);
|
|
||||||
expect(doTask).toBeCalledTimes(1);
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('test biz', () => {
|
|
||||||
let emitEvent: jest.SpyInstance;
|
|
||||||
beforeEach(() => {
|
|
||||||
emitEvent = jest
|
|
||||||
.spyOn(runner, 'emitEvent')
|
|
||||||
.mockImplementation(() => Promise.resolve());
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('doTask', () => {
|
|
||||||
let checkout: jest.SpyInstance;
|
|
||||||
let doTaskUnit: jest.SpyInstance;
|
|
||||||
|
|
||||||
beforeEach(() => {
|
|
||||||
checkout = jest
|
|
||||||
.spyOn(runner, 'checkout')
|
|
||||||
.mockImplementation(() => Promise.resolve('/null'));
|
|
||||||
doTaskUnit = jest
|
|
||||||
.spyOn(runner, 'doTaskUnit')
|
|
||||||
.mockImplementation(() => Promise.resolve());
|
|
||||||
});
|
|
||||||
|
|
||||||
it('only checkout', async () => {
|
|
||||||
const task = new PipelineTask();
|
|
||||||
(task.id = 'taskId'), (task.pipeline = new Pipeline());
|
|
||||||
task.units = [PipelineUnits.checkout];
|
|
||||||
task.pipeline.id = 'pipelineId';
|
|
||||||
task.pipeline.project = new Project();
|
|
||||||
task.pipeline.project.id = 'projectId';
|
|
||||||
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
|
|
||||||
task.pipeline.workUnitMetadata.version = 1;
|
|
||||||
task.pipeline.workUnitMetadata.units = [
|
|
||||||
{
|
|
||||||
type: PipelineUnits.checkout,
|
|
||||||
scripts: [],
|
|
||||||
},
|
|
||||||
];
|
|
||||||
|
|
||||||
await runner.doTask(task);
|
|
||||||
|
|
||||||
expect(checkout).toBeCalledTimes(1);
|
|
||||||
expect(doTaskUnit).toBeCalledTimes(0);
|
|
||||||
expect(emitEvent).toBeCalledTimes(2);
|
|
||||||
expect(emitEvent.mock.calls[0][0]).toMatchObject(task);
|
|
||||||
expect(emitEvent.mock.calls[0][1]).toBeNull();
|
|
||||||
expect(emitEvent.mock.calls[0][2]).toEqual(TaskStatuses.working);
|
|
||||||
expect(emitEvent.mock.calls[1][0]).toMatchObject(task);
|
|
||||||
expect(emitEvent.mock.calls[1][1]).toBeNull();
|
|
||||||
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.success);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('many units', async () => {
|
|
||||||
const task = new PipelineTask();
|
|
||||||
(task.id = 'taskId'), (task.pipeline = new Pipeline());
|
|
||||||
task.units = [
|
|
||||||
PipelineUnits.checkout,
|
|
||||||
PipelineUnits.test,
|
|
||||||
PipelineUnits.deploy,
|
|
||||||
];
|
|
||||||
task.pipeline.id = 'pipelineId';
|
|
||||||
task.pipeline.project = new Project();
|
|
||||||
task.pipeline.project.id = 'projectId';
|
|
||||||
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
|
|
||||||
task.pipeline.workUnitMetadata.version = 1;
|
|
||||||
task.pipeline.workUnitMetadata.units = [
|
|
||||||
{
|
|
||||||
type: PipelineUnits.checkout,
|
|
||||||
scripts: [],
|
|
||||||
},
|
|
||||||
{
|
|
||||||
type: PipelineUnits.installDependencies,
|
|
||||||
scripts: ['pwd'],
|
|
||||||
},
|
|
||||||
{
|
|
||||||
type: PipelineUnits.test,
|
|
||||||
scripts: ['pwd'],
|
|
||||||
},
|
|
||||||
{
|
|
||||||
type: PipelineUnits.deploy,
|
|
||||||
scripts: ['pwd', 'uname'],
|
|
||||||
},
|
|
||||||
];
|
|
||||||
|
|
||||||
await runner.doTask(task);
|
|
||||||
|
|
||||||
expect(checkout).toBeCalledTimes(1);
|
|
||||||
expect(doTaskUnit).toBeCalledTimes(1);
|
|
||||||
expect(emitEvent).toBeCalledTimes(2);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('unit work failed', async () => {
|
|
||||||
const task = new PipelineTask();
|
|
||||||
(task.id = 'taskId'), (task.pipeline = new Pipeline());
|
|
||||||
task.units = [PipelineUnits.checkout, PipelineUnits.test];
|
|
||||||
task.pipeline.id = 'pipelineId';
|
|
||||||
task.pipeline.project = new Project();
|
|
||||||
task.pipeline.project.id = 'projectId';
|
|
||||||
task.pipeline.workUnitMetadata = new WorkUnitMetadata();
|
|
||||||
task.pipeline.workUnitMetadata.version = 1;
|
|
||||||
task.pipeline.workUnitMetadata.units = [
|
|
||||||
{
|
|
||||||
type: PipelineUnits.checkout,
|
|
||||||
scripts: [],
|
|
||||||
},
|
|
||||||
{
|
|
||||||
type: PipelineUnits.test,
|
|
||||||
scripts: ['pwd'],
|
|
||||||
},
|
|
||||||
];
|
|
||||||
|
|
||||||
doTaskUnit = jest
|
|
||||||
.spyOn(runner, 'doTaskUnit')
|
|
||||||
.mockImplementation(() => Promise.reject(new Error('test error')));
|
|
||||||
await runner.doTask(task);
|
|
||||||
|
|
||||||
expect(checkout).toBeCalledTimes(1);
|
|
||||||
expect(doTaskUnit).toBeCalledTimes(1);
|
|
||||||
expect(emitEvent).toBeCalledTimes(2);
|
|
||||||
expect(emitEvent.mock.calls[1][0]).toMatchObject(task);
|
|
||||||
expect(emitEvent.mock.calls[1][1]).toBeNull();
|
|
||||||
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.failed);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('doTaskUnit', () => {
|
|
||||||
it('success', async () => {
|
|
||||||
const runScript = jest
|
|
||||||
.spyOn(runner, 'runScript')
|
|
||||||
.mockImplementation(() => Promise.resolve());
|
|
||||||
const task = new PipelineTask();
|
|
||||||
|
|
||||||
const unit = PipelineUnits.test;
|
|
||||||
const workspacePath = '/null';
|
|
||||||
await runner.doTaskUnit(unit, ['pwd'], task, workspacePath);
|
|
||||||
|
|
||||||
expect(emitEvent.mock.calls[0][0]).toEqual(task);
|
|
||||||
expect(emitEvent.mock.calls[0][1]).toEqual(unit);
|
|
||||||
expect(emitEvent.mock.calls[0][2]).toEqual(TaskStatuses.working);
|
|
||||||
expect(emitEvent.mock.calls[1][0]).toEqual(task);
|
|
||||||
expect(emitEvent.mock.calls[1][1]).toEqual(unit);
|
|
||||||
expect(emitEvent.mock.calls[1][2]).toEqual(TaskStatuses.success);
|
|
||||||
expect(runScript.mock.calls[0][0]).toEqual('pwd');
|
|
||||||
expect(runScript.mock.calls[0][1]).toEqual(workspacePath);
|
|
||||||
expect(runScript.mock.calls[0][2]).toEqual(task);
|
|
||||||
expect(runScript.mock.calls[0][3]).toEqual(unit);
|
|
||||||
});
|
|
||||||
it('failed', async () => {
|
|
||||||
const runScript = jest
|
|
||||||
.spyOn(runner, 'runScript')
|
|
||||||
.mockImplementation(() => Promise.reject(new Error('test error')));
|
|
||||||
const task = new PipelineTask();
|
|
||||||
|
|
||||||
const unit = PipelineUnits.test;
|
|
||||||
const workspacePath = '/null';
|
|
||||||
await expect(
|
|
||||||
runner.doTaskUnit(unit, ['pwd'], task, workspacePath),
|
|
||||||
).rejects.toThrow('test error');
|
|
||||||
|
|
||||||
expect(emitEvent.mock.calls[1]?.[0]).toEqual(task);
|
|
||||||
expect(emitEvent.mock.calls[1]?.[1]).toEqual(unit);
|
|
||||||
expect(emitEvent.mock.calls[1]?.[2]).toEqual(TaskStatuses.failed);
|
|
||||||
expect(runScript).toBeCalledTimes(1);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('runScript', () => {
|
|
||||||
it('normal', async () => {
|
|
||||||
const spawn = jest.fn<any, any>(() => ({
|
|
||||||
stdout: {
|
|
||||||
on: () => undefined,
|
|
||||||
},
|
|
||||||
stderr: {
|
|
||||||
on: () => undefined,
|
|
||||||
},
|
|
||||||
addListener: (_: any, fn: (code: number) => void) => {
|
|
||||||
fn(0);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
(runner as any).spawn = spawn;
|
|
||||||
|
|
||||||
const task = new PipelineTask();
|
|
||||||
task.id = 'taskId';
|
|
||||||
const unit = PipelineUnits.deploy;
|
|
||||||
|
|
||||||
await runner.runScript('script name', 'workspaceRoot', task, unit);
|
|
||||||
expect(spawn).toHaveBeenCalledTimes(1);
|
|
||||||
expect(spawn.mock.calls[0][0]).toEqual('script name');
|
|
||||||
expect(spawn.mock.calls[0][1]).toMatchObject({
|
|
||||||
shell: true,
|
|
||||||
cwd: 'workspaceRoot',
|
|
||||||
});
|
|
||||||
});
|
|
||||||
it('failed', async () => {
|
|
||||||
const spawn = jest.fn(() => ({
|
|
||||||
stdout: {
|
|
||||||
on: () => undefined,
|
|
||||||
},
|
|
||||||
stderr: {
|
|
||||||
on: () => undefined,
|
|
||||||
},
|
|
||||||
addListener: (_: any, fn: (code: number) => void) => {
|
|
||||||
fn(1);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
(runner as any).spawn = spawn;
|
|
||||||
|
|
||||||
const task = new PipelineTask();
|
|
||||||
task.id = 'taskId';
|
|
||||||
const unit = PipelineUnits.deploy;
|
|
||||||
|
|
||||||
expect(
|
|
||||||
runner.runScript('script name', 'workspaceRoot', task, unit),
|
|
||||||
).rejects.toThrowError();
|
|
||||||
});
|
|
||||||
it('wait emit message done', async () => {
|
|
||||||
let finishedFn: () => void;
|
|
||||||
const on = jest.fn((_: any, fn: (buff: Buffer) => void) => {
|
|
||||||
setTimeout(() => {
|
|
||||||
fn(Buffer.from('message 1'));
|
|
||||||
setTimeout(() => {
|
|
||||||
fn(Buffer.from('message 2'));
|
|
||||||
setTimeout(() => {
|
|
||||||
fn(Buffer.from('message 3'));
|
|
||||||
finishedFn();
|
|
||||||
}, 1000);
|
|
||||||
}, 10);
|
|
||||||
}, 10);
|
|
||||||
});
|
|
||||||
const spawn = jest.fn(() => ({
|
|
||||||
stdout: {
|
|
||||||
on,
|
|
||||||
},
|
|
||||||
stderr: {
|
|
||||||
on,
|
|
||||||
},
|
|
||||||
addListener: (_: any, fn: (code: number) => void) => {
|
|
||||||
finishedFn = () => fn(0);
|
|
||||||
},
|
|
||||||
}));
|
|
||||||
|
|
||||||
let emitSuccessCount = 0;
|
|
||||||
jest.spyOn(runner, 'emitEvent').mockImplementation(() => {
|
|
||||||
return new Promise((resolve) => {
|
|
||||||
setTimeout(() => {
|
|
||||||
emitSuccessCount++;
|
|
||||||
resolve();
|
|
||||||
}, 1000);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
(runner as any).spawn = spawn;
|
|
||||||
|
|
||||||
const task = new PipelineTask();
|
|
||||||
task.id = 'taskId';
|
|
||||||
const unit = PipelineUnits.deploy;
|
|
||||||
|
|
||||||
await runner.runScript('script name', 'workspaceRoot', task, unit);
|
|
||||||
expect(emitSuccessCount).toEqual(1 + 6);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('tryRunDeployScript', () => {
|
|
||||||
it('should be call deploy with right args', async () => {
|
|
||||||
const deploy = jest.spyOn(deployByPM2Service, 'deploy');
|
|
||||||
await expect(
|
|
||||||
runner['tryRunDeployScript'](
|
|
||||||
'/test/dir',
|
|
||||||
'@@DEPLOY ecosystem.config.js',
|
|
||||||
),
|
|
||||||
).resolves.toBe(true);
|
|
||||||
expect(deploy.mock.calls[0][0]).toEqual('/test/dir/ecosystem.config.js');
|
|
||||||
});
|
|
||||||
it('should return false', async () => {
|
|
||||||
await expect(
|
|
||||||
runner['tryRunDeployScript'](
|
|
||||||
'/test/dir',
|
|
||||||
'pm2 start ecosystem.config.js',
|
|
||||||
),
|
|
||||||
).resolves.toBe(false);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,339 +0,0 @@
|
|||||||
import { DeployByPm2Service } from './runners/deploy-by-pm2/deploy-by-pm2.service';
|
|
||||||
import { ReposService } from '../repos/repos.service';
|
|
||||||
import { spawn, ChildProcessWithoutNullStreams } from 'child_process';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
import { ApplicationException } from '../commons/exceptions/application.exception';
|
|
||||||
import { PipelineUnits } from './enums/pipeline-units.enum';
|
|
||||||
import { TaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
|
|
||||||
import {
|
|
||||||
AmqpConnection,
|
|
||||||
RabbitRPC,
|
|
||||||
RabbitSubscribe,
|
|
||||||
} from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
|
||||||
import { last } from 'ramda';
|
|
||||||
import { Inject } from '@nestjs/common';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
QUEUE_PIPELINE_TASK_KILL,
|
|
||||||
ROUTE_PIPELINE_TASK_KILL,
|
|
||||||
} from './pipeline-tasks.constants';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PIPELINE_TASK_FANOUT,
|
|
||||||
ROUTE_PIPELINE_TASK_LOG,
|
|
||||||
} from './pipeline-tasks.constants';
|
|
||||||
import {
|
|
||||||
getInstanceName,
|
|
||||||
getSelfInstanceQueueKey,
|
|
||||||
getSelfInstanceRouteKey,
|
|
||||||
} from '../commons/utils/rabbit-mq';
|
|
||||||
import { rm } from 'fs/promises';
|
|
||||||
import { rename } from 'fs/promises';
|
|
||||||
import { join } from 'path';
|
|
||||||
|
|
||||||
type Spawn = typeof spawn;
|
|
||||||
|
|
||||||
export class PipelineTaskRunner {
|
|
||||||
readonly processes = new Map<string, ChildProcessWithoutNullStreams>();
|
|
||||||
readonly stopTaskIds = new Set<string>();
|
|
||||||
|
|
||||||
constructor(
|
|
||||||
private readonly reposService: ReposService,
|
|
||||||
@InjectPinoLogger(PipelineTaskRunner.name)
|
|
||||||
private readonly logger: PinoLogger,
|
|
||||||
@Inject('spawn')
|
|
||||||
private readonly spawn: Spawn,
|
|
||||||
private readonly amqpConnection: AmqpConnection,
|
|
||||||
private readonly deployByPM2Service: DeployByPm2Service,
|
|
||||||
) {}
|
|
||||||
@RabbitSubscribe({
|
|
||||||
exchange: 'new-pipeline-task',
|
|
||||||
routingKey: 'mac',
|
|
||||||
queue: 'mac.new-pipeline-task',
|
|
||||||
})
|
|
||||||
async onNewTask(task: PipelineTask) {
|
|
||||||
this.logger.info({ task }, 'on new task [%s].', task.id);
|
|
||||||
try {
|
|
||||||
await this.doTask(task);
|
|
||||||
} catch (err) {
|
|
||||||
this.logger.error({ task, err }, err.message);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@RabbitRPC({
|
|
||||||
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
routingKey: getSelfInstanceRouteKey(ROUTE_PIPELINE_TASK_KILL),
|
|
||||||
queue: getSelfInstanceQueueKey(QUEUE_PIPELINE_TASK_KILL),
|
|
||||||
queueOptions: {
|
|
||||||
autoDelete: true,
|
|
||||||
durable: true,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async onStopTask(task: PipelineTask) {
|
|
||||||
this.logger.info({ task }, 'on stop task [%s].', task.id);
|
|
||||||
this.stopTaskIds.add(task.id);
|
|
||||||
const process = this.processes.get(task.id);
|
|
||||||
if (process) {
|
|
||||||
this.logger.info({ task }, 'send signal SIGINT to child process.');
|
|
||||||
process.kill('SIGINT');
|
|
||||||
|
|
||||||
setTimeout(() => {
|
|
||||||
setTimeout(() => {
|
|
||||||
this.stopTaskIds.delete(task.id);
|
|
||||||
}, 10_000);
|
|
||||||
if (process === this.processes.get(task.id)) {
|
|
||||||
this.logger.info({ task }, 'send signal SIGKILL to child process.');
|
|
||||||
process.kill('SIGKILL');
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
if (this.processes.has(task.id)) {
|
|
||||||
this.logger.error(
|
|
||||||
{ task },
|
|
||||||
'this pipeline task not stop yet. there is a new process running, maybe is a bug about error capture',
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}, 10_000);
|
|
||||||
} else {
|
|
||||||
this.logger.info({ task }, 'child process is not running.');
|
|
||||||
}
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
async doTask(task: PipelineTask) {
|
|
||||||
if (task.pipeline.workUnitMetadata.version !== 1) {
|
|
||||||
throw new ApplicationException(
|
|
||||||
'work unit metadata version is not match.',
|
|
||||||
);
|
|
||||||
}
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
null,
|
|
||||||
TaskStatuses.working,
|
|
||||||
`[start task]`,
|
|
||||||
'stdout',
|
|
||||||
);
|
|
||||||
|
|
||||||
this.logger.info('running task [%s].', task.id);
|
|
||||||
try {
|
|
||||||
let workspaceRoot = await this.checkout(task);
|
|
||||||
const units = task.units
|
|
||||||
.filter((unit) => unit !== PipelineUnits.checkout)
|
|
||||||
.map(
|
|
||||||
(type) =>
|
|
||||||
task.pipeline.workUnitMetadata.units.find(
|
|
||||||
(unit) => unit.type === type,
|
|
||||||
) ?? { type: type, scripts: [] },
|
|
||||||
);
|
|
||||||
this.logger.info({ units }, 'begin run units.');
|
|
||||||
for (const unit of units) {
|
|
||||||
if (unit.type === PipelineUnits.deploy) {
|
|
||||||
const oldRoot = workspaceRoot;
|
|
||||||
workspaceRoot = this.reposService.getDeployRoot(task);
|
|
||||||
if (oldRoot !== workspaceRoot) {
|
|
||||||
await rm(workspaceRoot, { force: true, recursive: true });
|
|
||||||
await rename(oldRoot, workspaceRoot);
|
|
||||||
}
|
|
||||||
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
unit.type,
|
|
||||||
TaskStatuses.success,
|
|
||||||
`[deploy] change deploy folder content success`,
|
|
||||||
'stdout',
|
|
||||||
);
|
|
||||||
}
|
|
||||||
await this.doTaskUnit(unit.type, unit.scripts, task, workspaceRoot);
|
|
||||||
}
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
null,
|
|
||||||
TaskStatuses.success,
|
|
||||||
`[finished task] success`,
|
|
||||||
'stdout',
|
|
||||||
);
|
|
||||||
this.logger.info({ task }, 'task [%s] completed.', task.id);
|
|
||||||
} catch (err) {
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
null,
|
|
||||||
TaskStatuses.failed,
|
|
||||||
`[finished unit] ${err.message}`,
|
|
||||||
'stderr',
|
|
||||||
);
|
|
||||||
this.logger.error({ task, error: err }, 'task [%s] failed.', task.id);
|
|
||||||
} finally {
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async doTaskUnit(
|
|
||||||
unit: PipelineUnits,
|
|
||||||
scripts: string[],
|
|
||||||
task: PipelineTask,
|
|
||||||
workspaceRoot: string,
|
|
||||||
) {
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
unit,
|
|
||||||
TaskStatuses.working,
|
|
||||||
`[begin unit] ${unit}`,
|
|
||||||
'stdin',
|
|
||||||
);
|
|
||||||
this.logger.info({ task }, 'curr unit is %s', unit);
|
|
||||||
try {
|
|
||||||
for (const script of scripts) {
|
|
||||||
this.logger.debug('begin runScript %s', script);
|
|
||||||
if (this.stopTaskIds.has(task.id)) {
|
|
||||||
throw new ApplicationException('Task is be KILLED');
|
|
||||||
}
|
|
||||||
await this.runScript(script, workspaceRoot, task, unit);
|
|
||||||
this.logger.debug('end runScript %s', script);
|
|
||||||
}
|
|
||||||
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
unit,
|
|
||||||
TaskStatuses.success,
|
|
||||||
`[finished unit] ${unit}`,
|
|
||||||
'stdout',
|
|
||||||
);
|
|
||||||
} catch (err) {
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
unit,
|
|
||||||
TaskStatuses.failed,
|
|
||||||
`[finished unit] ${err.message}`,
|
|
||||||
'stderr',
|
|
||||||
);
|
|
||||||
throw err;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async checkout(task: PipelineTask) {
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
PipelineUnits.checkout,
|
|
||||||
TaskStatuses.working,
|
|
||||||
'[begin unit] checkout',
|
|
||||||
'stdin',
|
|
||||||
);
|
|
||||||
try {
|
|
||||||
const path = await this.reposService.checkout4Task(task);
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
PipelineUnits.checkout,
|
|
||||||
TaskStatuses.success,
|
|
||||||
'checkout success.',
|
|
||||||
'stdout',
|
|
||||||
);
|
|
||||||
return path;
|
|
||||||
} catch (err) {
|
|
||||||
await this.emitEvent(
|
|
||||||
task,
|
|
||||||
PipelineUnits.checkout,
|
|
||||||
TaskStatuses.failed,
|
|
||||||
'checkout failed.',
|
|
||||||
'stderr',
|
|
||||||
);
|
|
||||||
throw err;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async emitEvent(
|
|
||||||
task: PipelineTask,
|
|
||||||
unit: PipelineUnits | null,
|
|
||||||
status: TaskStatuses,
|
|
||||||
message: string,
|
|
||||||
messageType: 'stderr' | 'stdout' | 'stdin',
|
|
||||||
) {
|
|
||||||
const event: PipelineTaskEvent = {
|
|
||||||
taskId: task.id,
|
|
||||||
pipelineId: task.pipeline.id,
|
|
||||||
projectId: task.pipeline.project.id,
|
|
||||||
unit,
|
|
||||||
emittedAt: new Date(),
|
|
||||||
message: last(message) === '\n' ? message : message + '\n',
|
|
||||||
messageType,
|
|
||||||
status,
|
|
||||||
};
|
|
||||||
this.amqpConnection
|
|
||||||
.publish(EXCHANGE_PIPELINE_TASK_FANOUT, ROUTE_PIPELINE_TASK_LOG, event, {
|
|
||||||
headers: {
|
|
||||||
sender: getInstanceName(),
|
|
||||||
},
|
|
||||||
})
|
|
||||||
.catch((error) => {
|
|
||||||
this.logger.error(
|
|
||||||
{ error, event },
|
|
||||||
'send event message to queue failed. %s',
|
|
||||||
error.message,
|
|
||||||
);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
async runScript(
|
|
||||||
script: string,
|
|
||||||
workspaceRoot: string,
|
|
||||||
task: PipelineTask,
|
|
||||||
unit: PipelineUnits,
|
|
||||||
): Promise<void> {
|
|
||||||
await this.emitEvent(task, unit, TaskStatuses.working, script, 'stdin');
|
|
||||||
if (await this.tryRunDeployScript(workspaceRoot, script)) {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
return new Promise((resolve, reject) => {
|
|
||||||
const sub = this.spawn(script, {
|
|
||||||
shell: true,
|
|
||||||
cwd: workspaceRoot,
|
|
||||||
});
|
|
||||||
this.processes.set(task.id, sub);
|
|
||||||
let loggingCount = 0; // semaphore
|
|
||||||
|
|
||||||
sub.stderr.on('data', (data: Buffer) => {
|
|
||||||
const str = data.toString();
|
|
||||||
loggingCount++;
|
|
||||||
|
|
||||||
this.emitEvent(task, unit, TaskStatuses.working, str, 'stdout').finally(
|
|
||||||
() => loggingCount--,
|
|
||||||
);
|
|
||||||
});
|
|
||||||
sub.stdout.on('data', (data: Buffer) => {
|
|
||||||
const str = data.toString();
|
|
||||||
loggingCount++;
|
|
||||||
|
|
||||||
this.emitEvent(task, unit, TaskStatuses.working, str, 'stderr').finally(
|
|
||||||
() => loggingCount--,
|
|
||||||
);
|
|
||||||
});
|
|
||||||
sub.addListener('close', async (code) => {
|
|
||||||
this.processes.delete(task.id);
|
|
||||||
await new Promise<void>(async (resolve) => {
|
|
||||||
for (let i = 0; i < 10 && loggingCount > 0; i++) {
|
|
||||||
await new Promise((resolve) => setTimeout(resolve, 500));
|
|
||||||
this.logger.debug('waiting logging... (%dx500ms)', i);
|
|
||||||
}
|
|
||||||
resolve();
|
|
||||||
});
|
|
||||||
if (code === 0) {
|
|
||||||
return resolve();
|
|
||||||
}
|
|
||||||
if (this.stopTaskIds.has(task.id)) {
|
|
||||||
throw reject(new ApplicationException('Task is be KILLED'));
|
|
||||||
}
|
|
||||||
return reject(new ApplicationException('exec script failed'));
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
private async tryRunDeployScript(workspaceRoot: string, script: string) {
|
|
||||||
const match = /^@@DEPLOY\s+(\S*)/.exec(script);
|
|
||||||
if (match) {
|
|
||||||
await this.deployByPM2Service.deploy(
|
|
||||||
join(workspaceRoot, match[1]),
|
|
||||||
workspaceRoot,
|
|
||||||
);
|
|
||||||
return true;
|
|
||||||
} else {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,11 +0,0 @@
|
|||||||
export const EXCHANGE_PIPELINE_TASK_TOPIC = 'pipeline-task.topic';
|
|
||||||
export const EXCHANGE_PIPELINE_TASK_FANOUT = 'pipeline-task.fanout';
|
|
||||||
export const ROUTE_PIPELINE_TASK_LOG = 'pipeline-task-log';
|
|
||||||
export const QUEUE_HANDLE_PIPELINE_TASK_LOG_EVENT = 'pipeline-task-log';
|
|
||||||
export const QUEUE_WRITE_PIPELINE_TASK_LOG = 'write-pipeline-task-log';
|
|
||||||
export const ROUTE_PIPELINE_TASK_DONE = 'pipeline-task-done';
|
|
||||||
export const QUEUE_PIPELINE_TASK_DONE = 'pipeline-task-done';
|
|
||||||
export const ROUTE_PIPELINE_TASK_KILL = 'pipeline-task-kill';
|
|
||||||
export const QUEUE_PIPELINE_TASK_KILL = 'pipeline-task-kill';
|
|
||||||
|
|
||||||
export const PM2_TOKEN = Symbol('pm2-token');
|
|
@ -1,76 +0,0 @@
|
|||||||
import { Module } from '@nestjs/common';
|
|
||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
|
||||||
import { PipelineTasksResolver } from './pipeline-tasks.resolver';
|
|
||||||
import { TypeOrmModule } from '@nestjs/typeorm';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
|
||||||
import { ReposModule } from '../repos/repos.module';
|
|
||||||
import { RedisModule } from 'nestjs-redis';
|
|
||||||
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { ConfigModule, ConfigService } from '@nestjs/config';
|
|
||||||
import { PipelineTaskRunner } from './pipeline-task.runner';
|
|
||||||
import { spawn } from 'child_process';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PIPELINE_TASK_FANOUT,
|
|
||||||
EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
} from './pipeline-tasks.constants';
|
|
||||||
import { PipelineTaskLogger } from './pipeline-task.logger';
|
|
||||||
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
|
|
||||||
import { CommonsModule } from '../commons/commons.module';
|
|
||||||
import { DeployByPm2Service } from './runners/deploy-by-pm2/deploy-by-pm2.service';
|
|
||||||
|
|
||||||
@Module({
|
|
||||||
imports: [
|
|
||||||
CommonsModule,
|
|
||||||
TypeOrmModule.forFeature([PipelineTask, Pipeline]),
|
|
||||||
RedisModule,
|
|
||||||
ReposModule,
|
|
||||||
RabbitMQModule.forRootAsync(RabbitMQModule, {
|
|
||||||
imports: [ConfigModule],
|
|
||||||
useFactory: (configService: ConfigService) => ({
|
|
||||||
uri: configService.get<string>('db.rabbitmq.uri'),
|
|
||||||
exchanges: [
|
|
||||||
{
|
|
||||||
name: 'new-pipeline-task',
|
|
||||||
type: 'fanout',
|
|
||||||
options: {
|
|
||||||
durable: true,
|
|
||||||
autoDelete: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: EXCHANGE_PIPELINE_TASK_FANOUT,
|
|
||||||
type: 'fanout',
|
|
||||||
options: {
|
|
||||||
durable: false,
|
|
||||||
autoDelete: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
name: EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
type: 'topic',
|
|
||||||
options: {
|
|
||||||
durable: false,
|
|
||||||
autoDelete: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}),
|
|
||||||
inject: [ConfigService],
|
|
||||||
}),
|
|
||||||
],
|
|
||||||
providers: [
|
|
||||||
PipelineTasksService,
|
|
||||||
PipelineTasksResolver,
|
|
||||||
PipelineTaskRunner,
|
|
||||||
PipelineTaskLogger,
|
|
||||||
{
|
|
||||||
provide: 'spawn',
|
|
||||||
useValue: spawn,
|
|
||||||
},
|
|
||||||
PipelineTaskFlushService,
|
|
||||||
DeployByPm2Service,
|
|
||||||
],
|
|
||||||
exports: [PipelineTasksService],
|
|
||||||
})
|
|
||||||
export class PipelineTasksModule {}
|
|
@ -1,35 +0,0 @@
|
|||||||
import { JwtService } from '@nestjs-lib/auth';
|
|
||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { PipelineTaskLogger } from './pipeline-task.logger';
|
|
||||||
import { PipelineTasksResolver } from './pipeline-tasks.resolver';
|
|
||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
|
||||||
|
|
||||||
describe('PipelineTasksResolver', () => {
|
|
||||||
let resolver: PipelineTasksResolver;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
PipelineTasksResolver,
|
|
||||||
{
|
|
||||||
provide: PipelineTasksService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: PipelineTaskLogger,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: JwtService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
resolver = module.get<PipelineTasksResolver>(PipelineTasksResolver);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(resolver).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,63 +0,0 @@
|
|||||||
import { Resolver, Args, Mutation, Subscription, Query } from '@nestjs/graphql';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
|
||||||
import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input';
|
|
||||||
import { PipelineTaskLogArgs } from './dtos/pipeline-task-log.args';
|
|
||||||
import { plainToClass } from 'class-transformer';
|
|
||||||
import { PipelineTaskLogger } from './pipeline-task.logger';
|
|
||||||
import { observableToAsyncIterable } from '@graphql-tools/utils';
|
|
||||||
import { PipelineTaskEvent } from './models/pipeline-task-event';
|
|
||||||
import { Roles, AccountRole } from '@nestjs-lib/auth';
|
|
||||||
|
|
||||||
@Roles(AccountRole.admin, AccountRole.super)
|
|
||||||
@Resolver()
|
|
||||||
export class PipelineTasksResolver {
|
|
||||||
constructor(
|
|
||||||
private readonly service: PipelineTasksService,
|
|
||||||
private readonly taskLogger: PipelineTaskLogger,
|
|
||||||
) {}
|
|
||||||
|
|
||||||
@Mutation(() => PipelineTask)
|
|
||||||
async createPipelineTask(@Args('task') taskDto: CreatePipelineTaskInput) {
|
|
||||||
return await this.service.addTask(taskDto);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Subscription(() => PipelineTaskEvent, {
|
|
||||||
resolve: (value) => {
|
|
||||||
const data = plainToClass(PipelineTaskEvent, value);
|
|
||||||
return data;
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async pipelineTaskEvent(@Args() args: PipelineTaskLogArgs) {
|
|
||||||
const task = await this.service.findTaskById(args.taskId);
|
|
||||||
return observableToAsyncIterable<PipelineTaskEvent>(
|
|
||||||
this.taskLogger.getMessage$(task.id),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Subscription(() => PipelineTask, {
|
|
||||||
resolve: (value) => {
|
|
||||||
return value;
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async pipelineTaskChanged(@Args('id') id: string) {
|
|
||||||
// return await this.service.watchTaskUpdated(id);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Query(() => [PipelineTask])
|
|
||||||
async listPipelineTaskByPipelineId(@Args('pipelineId') pipelineId: string) {
|
|
||||||
return await this.service.listTasksByPipelineId(pipelineId);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Query(() => PipelineTask)
|
|
||||||
async pipelineTask(@Args('id') id: string) {
|
|
||||||
return await this.service.findTaskById(id);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Mutation(() => Boolean)
|
|
||||||
async stopPipelineTask(@Args('id') id: string) {
|
|
||||||
const task = await this.service.findTaskById(id);
|
|
||||||
await this.service.stopTask(task);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,99 +0,0 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { PipelineTasksService } from './pipeline-tasks.service';
|
|
||||||
import { getRepositoryToken } from '@nestjs/typeorm';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
|
||||||
import { Repository } from 'typeorm';
|
|
||||||
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
|
|
||||||
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
|
|
||||||
|
|
||||||
describe('PipelineTasksService', () => {
|
|
||||||
let service: PipelineTasksService;
|
|
||||||
let module: TestingModule;
|
|
||||||
let taskRepository: Repository<PipelineTask>;
|
|
||||||
let pipelineRepository: Repository<Pipeline>;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
module = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
PipelineTasksService,
|
|
||||||
{
|
|
||||||
provide: getRepositoryToken(PipelineTask),
|
|
||||||
useValue: new Repository(),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: getRepositoryToken(Pipeline),
|
|
||||||
useValue: new Repository(),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: AmqpConnection,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: PipelineTaskFlushService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: getLoggerToken(PipelineTasksService.name),
|
|
||||||
useValue: new PinoLogger({}),
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
service = module.get<PipelineTasksService>(PipelineTasksService);
|
|
||||||
taskRepository = module.get(getRepositoryToken(PipelineTask));
|
|
||||||
pipelineRepository = module.get(getRepositoryToken(Pipeline));
|
|
||||||
jest
|
|
||||||
.spyOn(taskRepository, 'save')
|
|
||||||
.mockImplementation(async (data: any) => data);
|
|
||||||
jest
|
|
||||||
.spyOn(taskRepository, 'create')
|
|
||||||
.mockImplementation((data: any) => data);
|
|
||||||
jest.spyOn(taskRepository, 'findOne').mockImplementation(async () => null);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(service).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
// describe('addTask', () => {
|
|
||||||
// beforeEach(() => {
|
|
||||||
// jest
|
|
||||||
// .spyOn(pipelineRepository, 'findOneOrFail')
|
|
||||||
// .mockImplementation(async () => getBasePipeline());
|
|
||||||
// });
|
|
||||||
// it('pipeline not found', async () => {
|
|
||||||
// jest.spyOn(taskRepository, 'findOneOrFail').mockImplementation(() => {
|
|
||||||
// throw new EntityNotFoundError(Pipeline, {});
|
|
||||||
// });
|
|
||||||
// await expect(
|
|
||||||
// service.addTask({ pipelineId: 'test', commit: 'test', units: [] }),
|
|
||||||
// ).rejects;
|
|
||||||
// });
|
|
||||||
// it('create task on db', async () => {
|
|
||||||
// const save = jest
|
|
||||||
// .spyOn(taskRepository, 'save')
|
|
||||||
// .mockImplementation(async (data: any) => data);
|
|
||||||
// const findOne = jest.spyOn(taskRepository, 'findOne');
|
|
||||||
// await service.addTask({ pipelineId: 'test', commit: 'test', units: [] }),
|
|
||||||
// expect(save.mock.calls[0][0]).toMatchObject({
|
|
||||||
// pipelineId: 'test',
|
|
||||||
// commit: 'test',
|
|
||||||
// units: [],
|
|
||||||
// });
|
|
||||||
// expect(findOne).toBeCalled();
|
|
||||||
// });
|
|
||||||
// it('add task', async () => {
|
|
||||||
// const lpush = jest.spyOn(redisClient, 'lpush');
|
|
||||||
// await service.addTask({ pipelineId: 'test', commit: 'test', units: [] });
|
|
||||||
// expect(typeof lpush.mock.calls[0][1] === 'string').toBeTruthy();
|
|
||||||
// expect(JSON.parse(lpush.mock.calls[0][1] as string)).toMatchObject({
|
|
||||||
// pipelineId: 'test',
|
|
||||||
// commit: 'test',
|
|
||||||
// units: [],
|
|
||||||
// pipeline: getBasePipeline(),
|
|
||||||
// });
|
|
||||||
// });
|
|
||||||
// });
|
|
||||||
});
|
|
@ -1,158 +0,0 @@
|
|||||||
import { BadRequestException, Injectable } from '@nestjs/common';
|
|
||||||
import { InjectRepository } from '@nestjs/typeorm';
|
|
||||||
import { PipelineTask } from './pipeline-task.entity';
|
|
||||||
import { Repository } from 'typeorm';
|
|
||||||
import { CreatePipelineTaskInput } from './dtos/create-pipeline-task.input';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
|
||||||
import debug from 'debug';
|
|
||||||
import { AmqpConnection, RabbitRPC } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
QUEUE_PIPELINE_TASK_DONE,
|
|
||||||
ROUTE_PIPELINE_TASK_DONE,
|
|
||||||
} from './pipeline-tasks.constants';
|
|
||||||
import { PipelineTaskFlushService } from './pipeline-task-flush.service';
|
|
||||||
import { find, isNil, propEq } from 'ramda';
|
|
||||||
import { PipelineTaskLogs } from './models/pipeline-task-logs.model';
|
|
||||||
import { TaskStatuses, terminalTaskStatuses } from './enums/task-statuses.enum';
|
|
||||||
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
|
|
||||||
import { getAppInstanceRouteKey } from '../commons/utils/rabbit-mq';
|
|
||||||
import { ROUTE_PIPELINE_TASK_KILL } from './pipeline-tasks.constants';
|
|
||||||
|
|
||||||
const log = debug('fennec:pipeline-tasks:service');
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class PipelineTasksService {
|
|
||||||
constructor(
|
|
||||||
@InjectRepository(PipelineTask)
|
|
||||||
private readonly repository: Repository<PipelineTask>,
|
|
||||||
@InjectRepository(Pipeline)
|
|
||||||
private readonly pipelineRepository: Repository<Pipeline>,
|
|
||||||
private readonly amqpConnection: AmqpConnection,
|
|
||||||
private readonly eventFlushService: PipelineTaskFlushService,
|
|
||||||
@InjectPinoLogger(PipelineTasksService.name)
|
|
||||||
private readonly logger: PinoLogger,
|
|
||||||
) {}
|
|
||||||
async addTask(dto: CreatePipelineTaskInput) {
|
|
||||||
const pipeline = await this.pipelineRepository.findOneOrFail({
|
|
||||||
where: { id: dto.pipelineId },
|
|
||||||
relations: ['project'],
|
|
||||||
});
|
|
||||||
// const hasUnfinishedTask = await this.repository
|
|
||||||
// .findOne({
|
|
||||||
// pipelineId: dto.pipelineId,
|
|
||||||
// commit: dto.commit,
|
|
||||||
// status: In([TaskStatuses.pending, TaskStatuses.working]),
|
|
||||||
// })
|
|
||||||
// .then((val) => !isNil(val));
|
|
||||||
// if (hasUnfinishedTask) {
|
|
||||||
// throw new ConflictException(
|
|
||||||
// 'There are the same tasks among the unfinished tasks!',
|
|
||||||
// );
|
|
||||||
// }
|
|
||||||
const task = await this.repository.save(this.repository.create(dto));
|
|
||||||
task.pipeline = pipeline;
|
|
||||||
|
|
||||||
this.amqpConnection.publish('new-pipeline-task', 'mac', task);
|
|
||||||
return task;
|
|
||||||
}
|
|
||||||
|
|
||||||
async findTaskById(id: string) {
|
|
||||||
return await this.repository.findOneOrFail({ id });
|
|
||||||
}
|
|
||||||
|
|
||||||
async listTasksByPipelineId(pipelineId: string) {
|
|
||||||
return await this.repository.find({ pipelineId });
|
|
||||||
}
|
|
||||||
|
|
||||||
async listTasksByCommitHash(hash: string) {
|
|
||||||
return await this.repository.find({
|
|
||||||
where: { commit: hash },
|
|
||||||
order: { createdAt: 'DESC' },
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
getRedisTokens(pipeline: Pipeline): [string, string] {
|
|
||||||
return [`pipeline-${pipeline.id}:lck`, `pipeline-${pipeline.id}:tasks`];
|
|
||||||
}
|
|
||||||
|
|
||||||
@RabbitRPC({
|
|
||||||
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
routingKey: ROUTE_PIPELINE_TASK_DONE,
|
|
||||||
queue: QUEUE_PIPELINE_TASK_DONE,
|
|
||||||
queueOptions: {
|
|
||||||
autoDelete: true,
|
|
||||||
durable: true,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async updateByEvent({ taskId, runOn }: { taskId: string; runOn: string }) {
|
|
||||||
try {
|
|
||||||
const [events, task] = await Promise.all([
|
|
||||||
this.eventFlushService.read(taskId),
|
|
||||||
this.findTaskById(taskId),
|
|
||||||
]);
|
|
||||||
this.logger.info('[updateByEvent] start. taskId: %s', taskId);
|
|
||||||
|
|
||||||
for (const event of events) {
|
|
||||||
if (isNil(event.unit)) {
|
|
||||||
if (
|
|
||||||
event.status !== TaskStatuses.pending &&
|
|
||||||
task.status === TaskStatuses.pending
|
|
||||||
) {
|
|
||||||
task.startedAt = event.emittedAt;
|
|
||||||
} else if (terminalTaskStatuses.includes(event.status)) {
|
|
||||||
task.endedAt = event.emittedAt;
|
|
||||||
}
|
|
||||||
task.status = event.status;
|
|
||||||
} else {
|
|
||||||
let l: PipelineTaskLogs = find<PipelineTaskLogs>(
|
|
||||||
propEq('unit', event.unit),
|
|
||||||
task.logs,
|
|
||||||
);
|
|
||||||
|
|
||||||
if (isNil(l)) {
|
|
||||||
l = {
|
|
||||||
unit: event.unit,
|
|
||||||
startedAt: event.emittedAt,
|
|
||||||
endedAt: null,
|
|
||||||
logs: event.message,
|
|
||||||
status: event.status,
|
|
||||||
};
|
|
||||||
|
|
||||||
task.logs.push(l);
|
|
||||||
} else {
|
|
||||||
l.logs += event.message;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (terminalTaskStatuses.includes(event.status)) {
|
|
||||||
l.endedAt = event.emittedAt;
|
|
||||||
}
|
|
||||||
l.status = event.status;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
task.runOn = runOn;
|
|
||||||
await this.repository.update({ id: taskId }, task);
|
|
||||||
this.logger.info('[updateByEvent] success. taskId: %s', taskId);
|
|
||||||
return task;
|
|
||||||
} catch (error) {
|
|
||||||
this.logger.error(
|
|
||||||
{ error },
|
|
||||||
'[updateByEvent] failed. taskId: %s',
|
|
||||||
taskId,
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async stopTask(task: PipelineTask) {
|
|
||||||
if (isNil(task.runOn)) {
|
|
||||||
throw new BadRequestException(
|
|
||||||
"the task have not running instance on database. field 'runOn' is nil",
|
|
||||||
);
|
|
||||||
}
|
|
||||||
await this.amqpConnection.request({
|
|
||||||
exchange: EXCHANGE_PIPELINE_TASK_TOPIC,
|
|
||||||
routingKey: getAppInstanceRouteKey(ROUTE_PIPELINE_TASK_KILL, task.runOn),
|
|
||||||
payload: task,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,117 +0,0 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
|
|
||||||
import { join } from 'path';
|
|
||||||
import { DeployByPm2Service } from './deploy-by-pm2.service';
|
|
||||||
|
|
||||||
describe('DeployByPm2Service', () => {
|
|
||||||
let service: DeployByPm2Service;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
DeployByPm2Service,
|
|
||||||
{
|
|
||||||
provide: getLoggerToken(DeployByPm2Service.name),
|
|
||||||
useValue: new PinoLogger({
|
|
||||||
pinoHttp: {
|
|
||||||
level: 'silent',
|
|
||||||
},
|
|
||||||
}),
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
service = module.get<DeployByPm2Service>(DeployByPm2Service);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(service).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('getAppsSn', () => {
|
|
||||||
it('should return right value', () => {
|
|
||||||
expect(
|
|
||||||
service['getAppsSn']([
|
|
||||||
{ name: 'app' },
|
|
||||||
{ name: 'app#4' },
|
|
||||||
{ name: 'app#1' },
|
|
||||||
]),
|
|
||||||
).toEqual(4 + 1);
|
|
||||||
});
|
|
||||||
it('should return 1 when no match', () => {
|
|
||||||
expect(
|
|
||||||
service['getAppsSn']([
|
|
||||||
{ name: 'bar' },
|
|
||||||
{ name: 'foo#4' },
|
|
||||||
{ name: 'foo#1' },
|
|
||||||
]),
|
|
||||||
).toEqual(4 + 1);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('filterOldApps', () => {
|
|
||||||
it('should return right value', () => {
|
|
||||||
expect(
|
|
||||||
service['filterOldApps'](
|
|
||||||
[{ name: 'app' }],
|
|
||||||
[
|
|
||||||
{ name: 'app' },
|
|
||||||
{ name: 'app#4' },
|
|
||||||
{ name: 'foo#2' },
|
|
||||||
{ name: 'bar' },
|
|
||||||
],
|
|
||||||
),
|
|
||||||
).toEqual([{ name: 'app' }, { name: 'app#4' }]);
|
|
||||||
});
|
|
||||||
it('should return [] when no match', () => {
|
|
||||||
expect(
|
|
||||||
service['filterOldApps'](
|
|
||||||
[{ name: 'app' }],
|
|
||||||
[{ name: 'foo#2' }, { name: 'bar' }],
|
|
||||||
),
|
|
||||||
).toEqual([]);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('replaceAppName', () => {
|
|
||||||
it('should be replaced with right value', () => {
|
|
||||||
const getAppsSn = jest
|
|
||||||
.spyOn(service, 'getAppsSn' as any)
|
|
||||||
.mockImplementation(() => 1);
|
|
||||||
const options = [{ name: 'app' }, { name: 'foo' }];
|
|
||||||
service['replaceAppName'](options, []);
|
|
||||||
expect(options).toEqual([{ name: 'app#1' }, { name: 'foo#1' }]);
|
|
||||||
expect(getAppsSn).toBeCalledTimes(1);
|
|
||||||
expect(getAppsSn.mock.calls[0][0]).toEqual([]);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('deploy', () => {
|
|
||||||
it('should be success', async () => {
|
|
||||||
const filterOldApps = jest
|
|
||||||
.spyOn(service, 'filterOldApps' as any)
|
|
||||||
.mockImplementation(() => [{ name: 'app#1' }, { name: 'app#2' }]);
|
|
||||||
const replaceAppName = jest
|
|
||||||
.spyOn(service, 'replaceAppName' as any)
|
|
||||||
.mockImplementation((options) => (options[0].name = 'app#2'));
|
|
||||||
const stopApps = jest
|
|
||||||
.spyOn(service, 'stopApps' as any)
|
|
||||||
.mockImplementation(() => Promise.resolve());
|
|
||||||
await expect(
|
|
||||||
service['deploy'](
|
|
||||||
join(
|
|
||||||
__dirname,
|
|
||||||
'../../../../test/__mocks__/deploy-service/ecosystem.config.js',
|
|
||||||
),
|
|
||||||
join(__dirname, '../../../../test/__mocks__/deploy-service'),
|
|
||||||
),
|
|
||||||
).resolves.toBeFalsy();
|
|
||||||
expect(filterOldApps).toBeCalledTimes(1);
|
|
||||||
expect(replaceAppName).toBeCalledTimes(1);
|
|
||||||
expect(stopApps).toBeCalledTimes(1);
|
|
||||||
stopApps.mockReset();
|
|
||||||
|
|
||||||
await service['stopApps']([{ name: 'app#2' }]);
|
|
||||||
}, 10_000);
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,98 +0,0 @@
|
|||||||
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
|
|
||||||
import { Injectable } from '@nestjs/common';
|
|
||||||
import { promisify } from 'util';
|
|
||||||
import * as pm2 from 'pm2';
|
|
||||||
import { Proc, ProcessDescription, StartOptions } from 'pm2';
|
|
||||||
import { clone, last } from 'ramda';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class DeployByPm2Service {
|
|
||||||
constructor(
|
|
||||||
@InjectPinoLogger(DeployByPm2Service.name)
|
|
||||||
private readonly logger: PinoLogger,
|
|
||||||
) {}
|
|
||||||
async deploy(filePath: string, workspace: string) {
|
|
||||||
const baseConfig: { apps: StartOptions[] } = await import(filePath);
|
|
||||||
const appOptionsList: StartOptions[] = clone(baseConfig.apps);
|
|
||||||
|
|
||||||
await promisify<void>(pm2.connect.bind(pm2))();
|
|
||||||
const allApps = await promisify(pm2.list.bind(pm2))();
|
|
||||||
try {
|
|
||||||
if (!Array.isArray(baseConfig.apps)) {
|
|
||||||
this.logger.error(
|
|
||||||
'the "apps" in the PM2 ecosystem configuration is not array',
|
|
||||||
);
|
|
||||||
throw new Error('apps is not array');
|
|
||||||
}
|
|
||||||
|
|
||||||
const oldApps = this.filterOldApps(appOptionsList, allApps);
|
|
||||||
|
|
||||||
this.replaceAppName(appOptionsList, oldApps);
|
|
||||||
for (const appOptions of appOptionsList) {
|
|
||||||
const proc = await promisify<StartOptions, Proc>(pm2.start.bind(pm2))({
|
|
||||||
...appOptions,
|
|
||||||
cwd: workspace,
|
|
||||||
});
|
|
||||||
this.logger.info({ proc }, `start ${appOptions.name}`);
|
|
||||||
}
|
|
||||||
await this.stopApps(oldApps);
|
|
||||||
} catch (err) {
|
|
||||||
await this.stopApps(appOptionsList);
|
|
||||||
throw err;
|
|
||||||
} finally {
|
|
||||||
pm2.disconnect();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private async stopApps(apps: ProcessDescription[] | StartOptions[]) {
|
|
||||||
await Promise.all(
|
|
||||||
apps.map(async (app: ProcessDescription | StartOptions) => {
|
|
||||||
let procAtStop: ProcessDescription;
|
|
||||||
let procAtDelete: ProcessDescription;
|
|
||||||
try {
|
|
||||||
const idOrName = 'pm_id' in app ? app.pm_id : app.name;
|
|
||||||
procAtStop = await promisify(pm2.stop.bind(pm2))(idOrName);
|
|
||||||
procAtDelete = await promisify(pm2.delete.bind(pm2))(idOrName);
|
|
||||||
this.logger.info('stop & delete %s success', app.name);
|
|
||||||
} catch (error) {
|
|
||||||
this.logger.error(
|
|
||||||
{ error, procAtStop, procAtDelete },
|
|
||||||
'stop & delete %s error',
|
|
||||||
app.name,
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
private replaceAppName(
|
|
||||||
optionsList: StartOptions[],
|
|
||||||
oldApps: ProcessDescription[],
|
|
||||||
) {
|
|
||||||
const appSn = this.getAppsSn(oldApps);
|
|
||||||
|
|
||||||
optionsList.forEach((options) => {
|
|
||||||
if (!options.name) {
|
|
||||||
this.logger.error('please give a name for application');
|
|
||||||
throw new Error('app name is not given');
|
|
||||||
}
|
|
||||||
options.name = `${options.name}#${appSn}`;
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
private filterOldApps(
|
|
||||||
optionsList: StartOptions[],
|
|
||||||
apps: ProcessDescription[],
|
|
||||||
) {
|
|
||||||
return apps.filter((app) =>
|
|
||||||
optionsList.some((options) => app.name.split('#')[0] === options.name),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
private getAppsSn(oldApps: ProcessDescription[]) {
|
|
||||||
const appsSn: number[] = oldApps.map(
|
|
||||||
(app) => +(app.name.split('#')?.[1] ?? 0),
|
|
||||||
);
|
|
||||||
return (last(appsSn.sort()) ?? 0) + 1;
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,35 +0,0 @@
|
|||||||
import { JwtService } from '@nestjs-lib/auth';
|
|
||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { PipelineTasksService } from '../pipeline-tasks/pipeline-tasks.service';
|
|
||||||
import { CommitLogsResolver } from './commit-logs.resolver';
|
|
||||||
import { PipelinesService } from './pipelines.service';
|
|
||||||
|
|
||||||
describe('CommitLogsResolver', () => {
|
|
||||||
let resolver: CommitLogsResolver;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
CommitLogsResolver,
|
|
||||||
{
|
|
||||||
provide: PipelinesService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: PipelineTasksService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: JwtService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
resolver = module.get<CommitLogsResolver>(CommitLogsResolver);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(resolver).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,44 +0,0 @@
|
|||||||
import { Roles, AccountRole } from '@nestjs-lib/auth';
|
|
||||||
import { Query } from '@nestjs/graphql';
|
|
||||||
import {
|
|
||||||
Args,
|
|
||||||
Parent,
|
|
||||||
ResolveField,
|
|
||||||
Resolver,
|
|
||||||
Subscription,
|
|
||||||
} from '@nestjs/graphql';
|
|
||||||
import { PipelineTasksService } from '../pipeline-tasks/pipeline-tasks.service';
|
|
||||||
import { Commit, LogFields } from '../repos/dtos/log-list.model';
|
|
||||||
import { PipelinesService } from './pipelines.service';
|
|
||||||
|
|
||||||
@Roles(AccountRole.admin, AccountRole.super)
|
|
||||||
@Resolver(() => Commit)
|
|
||||||
export class CommitLogsResolver {
|
|
||||||
constructor(
|
|
||||||
private readonly service: PipelinesService,
|
|
||||||
private readonly taskServices: PipelineTasksService,
|
|
||||||
) {}
|
|
||||||
@Subscription(() => String, { resolve: (val) => val, nullable: true })
|
|
||||||
async syncCommits(
|
|
||||||
@Args('pipelineId', { type: () => String })
|
|
||||||
pipelineId: string,
|
|
||||||
@Args('appInstance', { type: () => String, nullable: true })
|
|
||||||
appInstance?: string,
|
|
||||||
) {
|
|
||||||
const pipeline = await this.service.findOneWithProject(pipelineId);
|
|
||||||
const syncCommitsPromise = this.service.syncCommits(pipeline, appInstance);
|
|
||||||
return (async function* () {
|
|
||||||
yield await syncCommitsPromise;
|
|
||||||
})();
|
|
||||||
}
|
|
||||||
@ResolveField()
|
|
||||||
async tasks(@Parent() commit: LogFields) {
|
|
||||||
return await this.taskServices.listTasksByCommitHash(commit.hash);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Query(() => [Commit], { nullable: true })
|
|
||||||
async commits(@Args('pipelineId', { type: () => String }) id: string) {
|
|
||||||
const pipeline = await this.service.findOneWithProject(id);
|
|
||||||
return await this.service.listCommits(pipeline);
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,31 +0,0 @@
|
|||||||
import { Type } from 'class-transformer';
|
|
||||||
import { InputType } from '@nestjs/graphql';
|
|
||||||
import { WorkUnitMetadata } from '../../pipeline-tasks/models/work-unit-metadata.model';
|
|
||||||
import {
|
|
||||||
IsInstance,
|
|
||||||
IsOptional,
|
|
||||||
IsString,
|
|
||||||
IsUUID,
|
|
||||||
MaxLength,
|
|
||||||
ValidateNested,
|
|
||||||
} from 'class-validator';
|
|
||||||
|
|
||||||
@InputType({ isAbstract: true })
|
|
||||||
export class CreatePipelineInput {
|
|
||||||
@IsUUID()
|
|
||||||
projectId: string;
|
|
||||||
|
|
||||||
@IsString()
|
|
||||||
@MaxLength(100)
|
|
||||||
branch: string;
|
|
||||||
|
|
||||||
@IsString()
|
|
||||||
@MaxLength(32)
|
|
||||||
name: string;
|
|
||||||
|
|
||||||
@Type(() => WorkUnitMetadata)
|
|
||||||
@IsOptional()
|
|
||||||
@ValidateNested()
|
|
||||||
@IsInstance(WorkUnitMetadata)
|
|
||||||
workUnitMetadata: WorkUnitMetadata;
|
|
||||||
}
|
|
@ -1,8 +0,0 @@
|
|||||||
import { ArgsType } from '@nestjs/graphql';
|
|
||||||
import { IsUUID } from 'class-validator';
|
|
||||||
|
|
||||||
@ArgsType()
|
|
||||||
export class ListPipelineArgs {
|
|
||||||
@IsUUID()
|
|
||||||
projectId?: string;
|
|
||||||
}
|
|
@ -1,9 +0,0 @@
|
|||||||
import { InputType, OmitType } from '@nestjs/graphql';
|
|
||||||
import { CreatePipelineInput } from './create-pipeline.input';
|
|
||||||
|
|
||||||
@InputType()
|
|
||||||
export class UpdatePipelineInput extends OmitType(CreatePipelineInput, [
|
|
||||||
'projectId',
|
|
||||||
]) {
|
|
||||||
id: string;
|
|
||||||
}
|
|
@ -1,23 +0,0 @@
|
|||||||
import { Column, Entity, ManyToOne } from 'typeorm';
|
|
||||||
import { AppBaseEntity } from '../commons/entities/app-base-entity';
|
|
||||||
import { Project } from '../projects/project.entity';
|
|
||||||
import { ObjectType } from '@nestjs/graphql';
|
|
||||||
import { WorkUnitMetadata } from '../pipeline-tasks/models/work-unit-metadata.model';
|
|
||||||
|
|
||||||
@ObjectType()
|
|
||||||
@Entity()
|
|
||||||
export class Pipeline extends AppBaseEntity {
|
|
||||||
@ManyToOne(() => Project)
|
|
||||||
project: Project;
|
|
||||||
@Column()
|
|
||||||
projectId: string;
|
|
||||||
|
|
||||||
@Column({ comment: 'eg: remotes/origin/master' })
|
|
||||||
branch: string;
|
|
||||||
|
|
||||||
@Column()
|
|
||||||
name: string;
|
|
||||||
|
|
||||||
@Column({ type: 'jsonb' })
|
|
||||||
workUnitMetadata: WorkUnitMetadata;
|
|
||||||
}
|
|
@ -1,29 +0,0 @@
|
|||||||
import { Module } from '@nestjs/common';
|
|
||||||
import { PipelinesResolver } from './pipelines.resolver';
|
|
||||||
import { PipelinesService } from './pipelines.service';
|
|
||||||
import { TypeOrmModule } from '@nestjs/typeorm';
|
|
||||||
import { Pipeline } from './pipeline.entity';
|
|
||||||
import { CommitLogsResolver } from './commit-logs.resolver';
|
|
||||||
import { PipelineTasksModule } from '../pipeline-tasks/pipeline-tasks.module';
|
|
||||||
import { ReposModule } from '../repos/repos.module';
|
|
||||||
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { ConfigModule, ConfigService } from '@nestjs/config';
|
|
||||||
import { CommonsModule } from '../commons/commons.module';
|
|
||||||
|
|
||||||
@Module({
|
|
||||||
imports: [
|
|
||||||
CommonsModule,
|
|
||||||
TypeOrmModule.forFeature([Pipeline]),
|
|
||||||
PipelineTasksModule,
|
|
||||||
RabbitMQModule.forRootAsync(RabbitMQModule, {
|
|
||||||
imports: [ConfigModule],
|
|
||||||
useFactory: (configService: ConfigService) => ({
|
|
||||||
uri: configService.get<string>('db.rabbitmq.uri'),
|
|
||||||
exchanges: [],
|
|
||||||
}),
|
|
||||||
inject: [ConfigService],
|
|
||||||
}),
|
|
||||||
],
|
|
||||||
providers: [PipelinesResolver, PipelinesService, CommitLogsResolver],
|
|
||||||
})
|
|
||||||
export class PipelinesModule {}
|
|
@ -1,30 +0,0 @@
|
|||||||
import { JwtService } from '@nestjs-lib/auth';
|
|
||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { PipelinesResolver } from './pipelines.resolver';
|
|
||||||
import { PipelinesService } from './pipelines.service';
|
|
||||||
|
|
||||||
describe('PipelinesResolver', () => {
|
|
||||||
let resolver: PipelinesResolver;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
PipelinesResolver,
|
|
||||||
{
|
|
||||||
provide: PipelinesService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: JwtService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
resolver = module.get<PipelinesResolver>(PipelinesResolver);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(resolver).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,45 +0,0 @@
|
|||||||
import { Args, Mutation, Query, Resolver } from '@nestjs/graphql';
|
|
||||||
import { CreatePipelineInput } from './dtos/create-pipeline.input';
|
|
||||||
import { UpdatePipelineInput } from './dtos/update-pipeline.input';
|
|
||||||
import { Pipeline } from './pipeline.entity';
|
|
||||||
import { PipelinesService } from './pipelines.service';
|
|
||||||
import { ListPipelineArgs } from './dtos/list-pipelines.args';
|
|
||||||
import { Roles, AccountRole } from '@nestjs-lib/auth';
|
|
||||||
|
|
||||||
@Roles(AccountRole.admin, AccountRole.super)
|
|
||||||
@Resolver()
|
|
||||||
export class PipelinesResolver {
|
|
||||||
constructor(private readonly service: PipelinesService) {}
|
|
||||||
@Query(() => [Pipeline])
|
|
||||||
async pipelines(@Args() dto: ListPipelineArgs) {
|
|
||||||
return await this.service.list(dto);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Query(() => Pipeline)
|
|
||||||
async pipeline(@Args('id', { type: () => String }) id: string) {
|
|
||||||
return await this.service.findOne(id);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Mutation(() => Pipeline)
|
|
||||||
async createPipeline(
|
|
||||||
@Args('pipeline', { type: () => CreatePipelineInput })
|
|
||||||
dto: CreatePipelineInput,
|
|
||||||
) {
|
|
||||||
return await this.service.create(dto);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Mutation(() => Pipeline)
|
|
||||||
async updatePipeline(
|
|
||||||
@Args('pipeline', { type: () => UpdatePipelineInput })
|
|
||||||
dto: UpdatePipelineInput,
|
|
||||||
) {
|
|
||||||
const tmp = await this.service.update(dto);
|
|
||||||
console.log(tmp);
|
|
||||||
return tmp;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Mutation(() => Number)
|
|
||||||
async deletePipeline(@Args('id', { type: () => String }) id: string) {
|
|
||||||
return await this.service.remove(id);
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,49 +0,0 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { PipelinesService } from './pipelines.service';
|
|
||||||
import { Pipeline } from './pipeline.entity';
|
|
||||||
import { getRepositoryToken } from '@nestjs/typeorm';
|
|
||||||
import { Repository } from 'typeorm';
|
|
||||||
import { Project } from '../projects/project.entity';
|
|
||||||
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
|
|
||||||
describe('PipelinesService', () => {
|
|
||||||
let service: PipelinesService;
|
|
||||||
let repository: Repository<Pipeline>;
|
|
||||||
let pipeline: Pipeline;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
pipeline = Object.assign(new Pipeline(), {
|
|
||||||
id: 'test-pipeline',
|
|
||||||
name: 'pipeline',
|
|
||||||
branch: 'master',
|
|
||||||
projectId: 'test-project',
|
|
||||||
project: Object.assign(new Project(), {
|
|
||||||
id: 'test-project',
|
|
||||||
name: 'project',
|
|
||||||
} as Project),
|
|
||||||
} as Pipeline);
|
|
||||||
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
PipelinesService,
|
|
||||||
{
|
|
||||||
provide: getRepositoryToken(Pipeline),
|
|
||||||
useValue: {
|
|
||||||
findOneOrFail: jest.fn().mockImplementation(() => pipeline),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: AmqpConnection,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
service = module.get<PipelinesService>(PipelinesService);
|
|
||||||
repository = module.get(getRepositoryToken(Pipeline));
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(service).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,78 +0,0 @@
|
|||||||
import { Injectable } from '@nestjs/common';
|
|
||||||
import { InjectRepository } from '@nestjs/typeorm';
|
|
||||||
import { Pipeline } from './pipeline.entity';
|
|
||||||
import { Repository } from 'typeorm';
|
|
||||||
import { BaseDbService } from '../commons/services/base-db.service';
|
|
||||||
import { CreatePipelineInput } from './dtos/create-pipeline.input';
|
|
||||||
import { UpdatePipelineInput } from './dtos/update-pipeline.input';
|
|
||||||
import { ListPipelineArgs } from './dtos/list-pipelines.args';
|
|
||||||
import {
|
|
||||||
EXCHANGE_REPO,
|
|
||||||
ROUTE_FETCH,
|
|
||||||
ROUTE_LIST_COMMITS,
|
|
||||||
} from '../repos/repos.constants';
|
|
||||||
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { Commit } from '../repos/dtos/log-list.model';
|
|
||||||
import { getAppInstanceRouteKey } from '../commons/utils/rabbit-mq';
|
|
||||||
import { ApplicationException } from '../commons/exceptions/application.exception';
|
|
||||||
import { plainToClass } from 'class-transformer';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class PipelinesService extends BaseDbService<Pipeline> {
|
|
||||||
readonly uniqueFields: Array<Array<keyof Pipeline>> = [['projectId', 'name']];
|
|
||||||
constructor(
|
|
||||||
@InjectRepository(Pipeline)
|
|
||||||
readonly repository: Repository<Pipeline>,
|
|
||||||
private readonly amqpConnection: AmqpConnection,
|
|
||||||
) {
|
|
||||||
super(repository);
|
|
||||||
}
|
|
||||||
async list(dto: ListPipelineArgs) {
|
|
||||||
return this.repository.find(dto);
|
|
||||||
}
|
|
||||||
|
|
||||||
async findOneWithProject(id: string) {
|
|
||||||
return await this.repository.findOne({
|
|
||||||
where: { id },
|
|
||||||
relations: ['project'],
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
async create(dto: CreatePipelineInput) {
|
|
||||||
await this.isDuplicateEntity(dto);
|
|
||||||
return await this.repository.save(this.repository.create(dto));
|
|
||||||
}
|
|
||||||
|
|
||||||
async update(dto: UpdatePipelineInput) {
|
|
||||||
const old = await this.findOne(dto.id);
|
|
||||||
await this.isDuplicateEntityForUpdate(old, dto);
|
|
||||||
return await this.repository.save(this.repository.merge(old, dto));
|
|
||||||
}
|
|
||||||
|
|
||||||
async remove(id: string) {
|
|
||||||
return (await this.repository.softDelete({ id })).affected;
|
|
||||||
}
|
|
||||||
async syncCommits(pipeline: Pipeline, appInstance?: string) {
|
|
||||||
return await this.amqpConnection.request<string | null>({
|
|
||||||
exchange: EXCHANGE_REPO,
|
|
||||||
routingKey: getAppInstanceRouteKey(ROUTE_FETCH, appInstance),
|
|
||||||
payload: pipeline,
|
|
||||||
timeout: 120_000,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
async listCommits(pipeline: Pipeline) {
|
|
||||||
return await this.amqpConnection
|
|
||||||
.request<[Error, Commit[]]>({
|
|
||||||
exchange: EXCHANGE_REPO,
|
|
||||||
routingKey: ROUTE_LIST_COMMITS,
|
|
||||||
payload: pipeline,
|
|
||||||
timeout: 30_000,
|
|
||||||
})
|
|
||||||
.then(([error, list]) => {
|
|
||||||
if (error) {
|
|
||||||
throw new ApplicationException(error);
|
|
||||||
}
|
|
||||||
return plainToClass(Commit, list);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
@ -3,7 +3,6 @@ import {
|
|||||||
IsOptional,
|
IsOptional,
|
||||||
IsString,
|
IsString,
|
||||||
IsUrl,
|
IsUrl,
|
||||||
Matches,
|
|
||||||
MaxLength,
|
MaxLength,
|
||||||
MinLength,
|
MinLength,
|
||||||
} from 'class-validator';
|
} from 'class-validator';
|
||||||
@ -20,12 +19,7 @@ export class CreateProjectInput {
|
|||||||
@MinLength(2)
|
@MinLength(2)
|
||||||
comment: string;
|
comment: string;
|
||||||
|
|
||||||
@Matches(
|
@IsUrl({ protocols: ['ssh'] })
|
||||||
/^(?:ssh:\/\/)?(?:[\w\d-_]+@)?(?:[\w\d-_]+\.)*\w{2,10}(?::\d{1,5})?(?:\/[\w\d-_.]+)*/,
|
|
||||||
{
|
|
||||||
message: 'wrong ssh url',
|
|
||||||
},
|
|
||||||
)
|
|
||||||
@MaxLength(256)
|
@MaxLength(256)
|
||||||
sshUrl: string;
|
sshUrl: string;
|
||||||
|
|
||||||
|
@ -1,9 +1,5 @@
|
|||||||
import { InputType } from '@nestjs/graphql';
|
import { InputType } from '@nestjs/graphql';
|
||||||
import { IsUUID } from 'class-validator';
|
|
||||||
import { CreateProjectInput } from './create-project.input';
|
import { CreateProjectInput } from './create-project.input';
|
||||||
|
|
||||||
@InputType()
|
@InputType()
|
||||||
export class UpdateProjectInput extends CreateProjectInput {
|
export class UpdateProjectInput extends CreateProjectInput {}
|
||||||
@IsUUID()
|
|
||||||
id: string;
|
|
||||||
}
|
|
||||||
|
@ -22,4 +22,7 @@ export class Project extends AppBaseEntity {
|
|||||||
|
|
||||||
@Column({ nullable: true })
|
@Column({ nullable: true })
|
||||||
webHookSecret?: string;
|
webHookSecret?: string;
|
||||||
|
|
||||||
|
@DeleteDateColumn()
|
||||||
|
deletedAt?: Date;
|
||||||
}
|
}
|
||||||
|
@ -1,3 +0,0 @@
|
|||||||
export const EXCHANGE_PROJECT_TOPIC = 'project.topic';
|
|
||||||
export const EXCHANGE_PROJECT_FANOUT = 'project.fanout';
|
|
||||||
export const ROUTE_PROJECT_CHANGE = 'project-change';
|
|
@ -3,33 +3,9 @@ import { ProjectsService } from './projects.service';
|
|||||||
import { ProjectsResolver } from './projects.resolver';
|
import { ProjectsResolver } from './projects.resolver';
|
||||||
import { TypeOrmModule } from '@nestjs/typeorm';
|
import { TypeOrmModule } from '@nestjs/typeorm';
|
||||||
import { Project } from './project.entity';
|
import { Project } from './project.entity';
|
||||||
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { ConfigModule, ConfigService } from '@nestjs/config';
|
|
||||||
import { EXCHANGE_PROJECT_FANOUT } from './projects.constants';
|
|
||||||
import { CommonsModule } from '../commons/commons.module';
|
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [
|
imports: [TypeOrmModule.forFeature([Project])],
|
||||||
CommonsModule,
|
|
||||||
TypeOrmModule.forFeature([Project]),
|
|
||||||
RabbitMQModule.forRootAsync(RabbitMQModule, {
|
|
||||||
imports: [ConfigModule],
|
|
||||||
useFactory: (configService: ConfigService) => ({
|
|
||||||
uri: configService.get<string>('db.rabbitmq.uri'),
|
|
||||||
exchanges: [
|
|
||||||
{
|
|
||||||
name: EXCHANGE_PROJECT_FANOUT,
|
|
||||||
type: 'fanout',
|
|
||||||
options: {
|
|
||||||
durable: false,
|
|
||||||
autoDelete: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}),
|
|
||||||
inject: [ConfigService],
|
|
||||||
}),
|
|
||||||
],
|
|
||||||
providers: [ProjectsService, ProjectsResolver],
|
providers: [ProjectsService, ProjectsResolver],
|
||||||
exports: [ProjectsService],
|
exports: [ProjectsService],
|
||||||
})
|
})
|
||||||
|
@ -1,4 +1,3 @@
|
|||||||
import { JwtService } from '@nestjs-lib/auth';
|
|
||||||
import { Test, TestingModule } from '@nestjs/testing';
|
import { Test, TestingModule } from '@nestjs/testing';
|
||||||
import { ProjectsResolver } from './projects.resolver';
|
import { ProjectsResolver } from './projects.resolver';
|
||||||
import { ProjectsService } from './projects.service';
|
import { ProjectsService } from './projects.service';
|
||||||
@ -14,10 +13,6 @@ describe('ProjectsResolver', () => {
|
|||||||
provide: ProjectsService,
|
provide: ProjectsService,
|
||||||
useValue: {},
|
useValue: {},
|
||||||
},
|
},
|
||||||
{
|
|
||||||
provide: JwtService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
|
@ -1,44 +1,43 @@
|
|||||||
import { AccountRole, Roles } from '@nestjs-lib/auth';
|
|
||||||
import { Args, Mutation, Query, Resolver } from '@nestjs/graphql';
|
import { Args, Mutation, Query, Resolver } from '@nestjs/graphql';
|
||||||
import { CreateProjectInput } from './dtos/create-project.input';
|
import { CreateProjectInput } from './dtos/create-project.input';
|
||||||
import { UpdateProjectInput } from './dtos/update-project.input';
|
import { UpdateProjectInput } from './dtos/update-project.input';
|
||||||
import { Project } from './project.entity';
|
import { Project } from './project.entity';
|
||||||
import { ProjectsService } from './projects.service';
|
import { ProjectsService } from './projects.service';
|
||||||
|
|
||||||
@Roles(AccountRole.admin, AccountRole.super)
|
|
||||||
@Resolver(() => Project)
|
@Resolver(() => Project)
|
||||||
export class ProjectsResolver {
|
export class ProjectsResolver {
|
||||||
constructor(private readonly service: ProjectsService) {}
|
constructor(private readonly service: ProjectsService) {}
|
||||||
@Query(() => [Project])
|
@Query(() => [Project])
|
||||||
async projects() {
|
async findProjects() {
|
||||||
return await this.service.list();
|
return await this.service.list();
|
||||||
}
|
}
|
||||||
|
|
||||||
@Query(() => Project)
|
@Query(() => Project)
|
||||||
async project(@Args('id', { type: () => String }) id: string) {
|
async findProject(@Args('id', { type: () => String }) id: string) {
|
||||||
return await this.service.findOne(id);
|
return await this.service.findOne(id);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Mutation(() => Project)
|
@Mutation(() => Project)
|
||||||
async createProject(
|
async createProject(
|
||||||
@Args('project', { type: () => CreateProjectInput })
|
@Args('project', { type: () => CreateProjectInput })
|
||||||
dto: CreateProjectInput,
|
dto: UpdateProjectInput,
|
||||||
) {
|
) {
|
||||||
return await this.service.create(dto);
|
return await this.service.create(dto);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Mutation(() => Project)
|
@Mutation(() => Project)
|
||||||
async updateProject(
|
async modifyProject(
|
||||||
|
@Args('id', { type: () => String }) id: string,
|
||||||
@Args('project', { type: () => UpdateProjectInput })
|
@Args('project', { type: () => UpdateProjectInput })
|
||||||
dto: UpdateProjectInput,
|
dto: UpdateProjectInput,
|
||||||
) {
|
) {
|
||||||
const tmp = await this.service.update(dto);
|
const tmp = await this.service.update(id, dto);
|
||||||
console.log(tmp);
|
console.log(tmp);
|
||||||
return tmp;
|
return tmp;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Mutation(() => Number)
|
@Mutation(() => Number)
|
||||||
async removeProject(@Args('id', { type: () => String }) id: string) {
|
async deleteProject(@Args('id', { type: () => String }) id: string) {
|
||||||
return await this.service.remove(id);
|
return await this.service.remove(id);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -2,7 +2,6 @@ import { Test, TestingModule } from '@nestjs/testing';
|
|||||||
import { ProjectsService } from './projects.service';
|
import { ProjectsService } from './projects.service';
|
||||||
import { getRepositoryToken } from '@nestjs/typeorm';
|
import { getRepositoryToken } from '@nestjs/typeorm';
|
||||||
import { Project } from './project.entity';
|
import { Project } from './project.entity';
|
||||||
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
|
|
||||||
describe('ProjectsService', () => {
|
describe('ProjectsService', () => {
|
||||||
let service: ProjectsService;
|
let service: ProjectsService;
|
||||||
@ -15,10 +14,6 @@ describe('ProjectsService', () => {
|
|||||||
provide: getRepositoryToken(Project),
|
provide: getRepositoryToken(Project),
|
||||||
useValue: {},
|
useValue: {},
|
||||||
},
|
},
|
||||||
{
|
|
||||||
provide: AmqpConnection,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
|
@ -4,12 +4,6 @@ import { BaseDbService } from '../commons/services/base-db.service';
|
|||||||
import { Repository } from 'typeorm';
|
import { Repository } from 'typeorm';
|
||||||
import { CreateProjectInput } from './dtos/create-project.input';
|
import { CreateProjectInput } from './dtos/create-project.input';
|
||||||
import { Project } from './project.entity';
|
import { Project } from './project.entity';
|
||||||
import { UpdateProjectInput } from './dtos/update-project.input';
|
|
||||||
import { AmqpConnection } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PROJECT_FANOUT,
|
|
||||||
ROUTE_PROJECT_CHANGE,
|
|
||||||
} from './projects.constants';
|
|
||||||
|
|
||||||
@Injectable()
|
@Injectable()
|
||||||
export class ProjectsService extends BaseDbService<Project> {
|
export class ProjectsService extends BaseDbService<Project> {
|
||||||
@ -17,7 +11,6 @@ export class ProjectsService extends BaseDbService<Project> {
|
|||||||
constructor(
|
constructor(
|
||||||
@InjectRepository(Project)
|
@InjectRepository(Project)
|
||||||
readonly repository: Repository<Project>,
|
readonly repository: Repository<Project>,
|
||||||
private readonly amqpConnection: AmqpConnection,
|
|
||||||
) {
|
) {
|
||||||
super(repository);
|
super(repository);
|
||||||
}
|
}
|
||||||
@ -31,15 +24,10 @@ export class ProjectsService extends BaseDbService<Project> {
|
|||||||
return await this.repository.save(this.repository.create(dto));
|
return await this.repository.save(this.repository.create(dto));
|
||||||
}
|
}
|
||||||
|
|
||||||
async update(dto: UpdateProjectInput) {
|
async update(id: string, dto: CreateProjectInput) {
|
||||||
await this.isDuplicateEntityForUpdate(dto.id, dto);
|
await this.isDuplicateEntityForUpdate(id, dto);
|
||||||
const old = await this.findOne(dto.id);
|
const old = await this.findOne(id);
|
||||||
const project = await this.repository.save(this.repository.merge(old, dto));
|
return await this.repository.save(this.repository.merge(old, dto));
|
||||||
this.amqpConnection.publish(EXCHANGE_PROJECT_FANOUT, ROUTE_PROJECT_CHANGE, [
|
|
||||||
project,
|
|
||||||
old,
|
|
||||||
]);
|
|
||||||
return project;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async remove(id: string) {
|
async remove(id: string) {
|
||||||
|
@ -1,16 +0,0 @@
|
|||||||
import { InputType } from '@nestjs/graphql';
|
|
||||||
import { IsOptional, IsString, IsUUID } from 'class-validator';
|
|
||||||
|
|
||||||
@InputType()
|
|
||||||
export class CheckoutInput {
|
|
||||||
@IsUUID()
|
|
||||||
pipelineId: string;
|
|
||||||
|
|
||||||
@IsString()
|
|
||||||
@IsOptional()
|
|
||||||
branch?: string;
|
|
||||||
|
|
||||||
@IsString()
|
|
||||||
@IsOptional()
|
|
||||||
commitNumber?: string;
|
|
||||||
}
|
|
@ -1,4 +1,4 @@
|
|||||||
import { InputType } from '@nestjs/graphql';
|
import { InputType, ObjectType } from '@nestjs/graphql';
|
||||||
import { IsOptional, IsString, IsUUID } from 'class-validator';
|
import { IsOptional, IsString, IsUUID } from 'class-validator';
|
||||||
|
|
||||||
@InputType()
|
@InputType()
|
||||||
|
@ -1,20 +1,6 @@
|
|||||||
import { ObjectType, Field } from '@nestjs/graphql';
|
import { ObjectType, Field } from '@nestjs/graphql';
|
||||||
import { Type } from 'class-transformer';
|
|
||||||
import { LogResult, DefaultLogFields } from 'simple-git';
|
import { LogResult, DefaultLogFields } from 'simple-git';
|
||||||
import { PipelineTask } from '../../pipeline-tasks/pipeline-task.entity';
|
|
||||||
|
|
||||||
@ObjectType()
|
|
||||||
export class Commit {
|
|
||||||
hash: string;
|
|
||||||
@Type(() => Date)
|
|
||||||
date: Date;
|
|
||||||
message: string;
|
|
||||||
refs: string;
|
|
||||||
body: string;
|
|
||||||
author_name: string;
|
|
||||||
author_email: string;
|
|
||||||
tasks: PipelineTask[];
|
|
||||||
}
|
|
||||||
@ObjectType()
|
@ObjectType()
|
||||||
export class LogFields {
|
export class LogFields {
|
||||||
hash: string;
|
hash: string;
|
||||||
@ -24,7 +10,6 @@ export class LogFields {
|
|||||||
body: string;
|
body: string;
|
||||||
author_name: string;
|
author_name: string;
|
||||||
author_email: string;
|
author_email: string;
|
||||||
tasks: PipelineTask[];
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@ObjectType()
|
@ObjectType()
|
||||||
|
@ -1,5 +0,0 @@
|
|||||||
import { Project } from '../../projects/project.entity';
|
|
||||||
export interface ListLogsOption {
|
|
||||||
project: Project;
|
|
||||||
branch: string;
|
|
||||||
}
|
|
@ -1,6 +1 @@
|
|||||||
export const EXCHANGE_REPO = 'fennec.repo';
|
export const WORKSPACE_ACTION = 'workspace-action';
|
||||||
export const ROUTE_FETCH = 'fetch';
|
|
||||||
export const ROUTE_LIST_COMMITS = 'list-commits';
|
|
||||||
export const QUEUE_LIST_COMMITS = 'list-commits';
|
|
||||||
export const QUEUE_FETCH = 'repo-fetch';
|
|
||||||
export const QUEUE_REFRESH_REPO = 'refresh-repo';
|
|
||||||
|
@ -3,37 +3,18 @@ import { TypeOrmModule } from '@nestjs/typeorm';
|
|||||||
import { Project } from '../projects/project.entity';
|
import { Project } from '../projects/project.entity';
|
||||||
import { ReposResolver } from './repos.resolver';
|
import { ReposResolver } from './repos.resolver';
|
||||||
import { ReposService } from './repos.service';
|
import { ReposService } from './repos.service';
|
||||||
import { ConfigModule, ConfigService } from '@nestjs/config';
|
import { ConfigModule } from '@nestjs/config';
|
||||||
import { ProjectsModule } from '../projects/projects.module';
|
import { BullModule } from '@nestjs/bull';
|
||||||
import { EXCHANGE_REPO } from './repos.constants';
|
import { WORKSPACE_ACTION } from './repos.constants';
|
||||||
import { RabbitMQModule } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { CommonsModule } from '../commons/commons.module';
|
|
||||||
|
|
||||||
@Module({
|
@Module({
|
||||||
imports: [
|
imports: [
|
||||||
TypeOrmModule.forFeature([Project]),
|
TypeOrmModule.forFeature([Project]),
|
||||||
ConfigModule,
|
ConfigModule,
|
||||||
ProjectsModule,
|
BullModule.registerQueue({
|
||||||
CommonsModule,
|
name: WORKSPACE_ACTION,
|
||||||
RabbitMQModule.forRootAsync(RabbitMQModule, {
|
|
||||||
imports: [ConfigModule],
|
|
||||||
useFactory: (configService: ConfigService) => ({
|
|
||||||
uri: configService.get<string>('db.rabbitmq.uri'),
|
|
||||||
exchanges: [
|
|
||||||
{
|
|
||||||
name: EXCHANGE_REPO,
|
|
||||||
type: 'topic',
|
|
||||||
options: {
|
|
||||||
durable: true,
|
|
||||||
autoDelete: true,
|
|
||||||
},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}),
|
|
||||||
inject: [ConfigService],
|
|
||||||
}),
|
}),
|
||||||
],
|
],
|
||||||
providers: [ReposResolver, ReposService],
|
providers: [ReposResolver, ReposService],
|
||||||
exports: [ReposService],
|
|
||||||
})
|
})
|
||||||
export class ReposModule {}
|
export class ReposModule {}
|
||||||
|
@ -1,7 +1,6 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
import { Test, TestingModule } from '@nestjs/testing';
|
||||||
import { ReposResolver } from './repos.resolver';
|
import { ReposResolver } from './repos.resolver';
|
||||||
import { ReposService } from './repos.service';
|
import { ReposService } from './repos.service';
|
||||||
import { ProjectsService } from '../projects/projects.service';
|
|
||||||
|
|
||||||
describe('ReposResolver', () => {
|
describe('ReposResolver', () => {
|
||||||
let resolver: ReposResolver;
|
let resolver: ReposResolver;
|
||||||
@ -14,10 +13,6 @@ describe('ReposResolver', () => {
|
|||||||
provide: ReposService,
|
provide: ReposService,
|
||||||
useValue: {},
|
useValue: {},
|
||||||
},
|
},
|
||||||
{
|
|
||||||
provide: ProjectsService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
|
@ -1,4 +1,26 @@
|
|||||||
import { Resolver } from '@nestjs/graphql';
|
import { Args, Query, Resolver } from '@nestjs/graphql';
|
||||||
|
import { ListLogsArgs } from './dtos/list-logs.args';
|
||||||
|
import { ReposService } from './repos.service';
|
||||||
|
import { LogList } from './dtos/log-list.model';
|
||||||
|
import { ListBranchesArgs } from './dtos/list-branches.args';
|
||||||
|
import { BranchList } from './dtos/branch-list.model';
|
||||||
|
|
||||||
@Resolver()
|
@Resolver()
|
||||||
export class ReposResolver {}
|
export class ReposResolver {
|
||||||
|
constructor(private readonly service: ReposService) {}
|
||||||
|
@Query(() => LogList)
|
||||||
|
async listLogs(@Args('listLogsArgs') dto: ListLogsArgs) {
|
||||||
|
return await this.service.listLogs(dto);
|
||||||
|
}
|
||||||
|
@Query(() => BranchList)
|
||||||
|
async ListBranchesArgs(
|
||||||
|
@Args('listBranchesArgs') dto: ListBranchesArgs,
|
||||||
|
): Promise<BranchList> {
|
||||||
|
return await this.service.listBranches(dto).then((data) => {
|
||||||
|
return {
|
||||||
|
...data,
|
||||||
|
branches: Object.values(data.branches),
|
||||||
|
};
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@ -1,18 +1,12 @@
|
|||||||
import { Pipeline } from './../pipelines/pipeline.entity';
|
|
||||||
import { Test, TestingModule } from '@nestjs/testing';
|
import { Test, TestingModule } from '@nestjs/testing';
|
||||||
import { getRepositoryToken } from '@nestjs/typeorm';
|
import { getRepositoryToken } from '@nestjs/typeorm';
|
||||||
import { Project } from '../projects/project.entity';
|
import { Project } from '../projects/project.entity';
|
||||||
import { ReposService } from './repos.service';
|
import { ReposService } from './repos.service';
|
||||||
import { ConfigModule } from '@nestjs/config';
|
import { ConfigModule } from '@nestjs/config';
|
||||||
import { rm } from 'fs/promises';
|
import { readFile, rm } from 'fs/promises';
|
||||||
import configuration from '../commons/config/configuration';
|
|
||||||
import { PipelineTask } from '../pipeline-tasks/pipeline-task.entity';
|
|
||||||
import { join } from 'path';
|
import { join } from 'path';
|
||||||
import { readFile } from 'fs/promises';
|
import configuration from '../commons/config/configuration';
|
||||||
import { getLoggerToken, PinoLogger } from 'nestjs-pino';
|
import { NotFoundException } from '@nestjs/common';
|
||||||
import { Nack } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { getInstanceName } from '../commons/utils/rabbit-mq';
|
|
||||||
import { RedisMutexService } from '../commons/redis-mutex/redis-mutex.service';
|
|
||||||
|
|
||||||
const getTest1Project = () =>
|
const getTest1Project = () =>
|
||||||
({
|
({
|
||||||
@ -49,18 +43,6 @@ describe('ReposService', () => {
|
|||||||
provide: getRepositoryToken(Project),
|
provide: getRepositoryToken(Project),
|
||||||
useFactory: repositoryMockFactory,
|
useFactory: repositoryMockFactory,
|
||||||
},
|
},
|
||||||
{
|
|
||||||
provide: getLoggerToken(ReposService.name),
|
|
||||||
useValue: new PinoLogger({}),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: RedisMutexService,
|
|
||||||
useValue: {
|
|
||||||
lock: jest.fn(() =>
|
|
||||||
Promise.resolve(() => Promise.resolve(undefined)),
|
|
||||||
),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
],
|
],
|
||||||
}).compile();
|
}).compile();
|
||||||
|
|
||||||
@ -79,10 +61,7 @@ describe('ReposService', () => {
|
|||||||
});
|
});
|
||||||
describe('listLogs', () => {
|
describe('listLogs', () => {
|
||||||
it('should be return logs', async () => {
|
it('should be return logs', async () => {
|
||||||
const result = await service.listLogs({
|
const result = await service.listLogs({ projectId: '1' });
|
||||||
project: getTest1Project(),
|
|
||||||
branch: 'master',
|
|
||||||
});
|
|
||||||
expect(result).toBeDefined();
|
expect(result).toBeDefined();
|
||||||
}, 20_000);
|
}, 20_000);
|
||||||
});
|
});
|
||||||
@ -90,122 +69,63 @@ describe('ReposService', () => {
|
|||||||
it('should be return branches', async () => {
|
it('should be return branches', async () => {
|
||||||
const result = await service.listBranches({ projectId: '1' });
|
const result = await service.listBranches({ projectId: '1' });
|
||||||
expect(result).toBeDefined();
|
expect(result).toBeDefined();
|
||||||
}, 20_000);
|
}, 10_000);
|
||||||
|
});
|
||||||
|
describe('checkoutBranch', () => {
|
||||||
|
it('should be checkout', async () => {
|
||||||
|
await service.checkoutBranch(getTest1Project(), 'master');
|
||||||
|
const filePath = join(
|
||||||
|
service.getWorkspaceRoot(getTest1Project()),
|
||||||
|
'README.md',
|
||||||
|
);
|
||||||
|
const text = await readFile(filePath, { encoding: 'utf-8' });
|
||||||
|
expect(text).toMatch(/Commit 1/gi);
|
||||||
|
}, 30_000);
|
||||||
|
it('multiplexing workspace', async () => {
|
||||||
|
await service.checkoutBranch(getTest1Project(), 'master');
|
||||||
|
await service.checkoutBranch(getTest1Project(), 'branch-a');
|
||||||
|
await service.checkoutBranch(getTest1Project(), 'branch-b');
|
||||||
|
const filePath = join(
|
||||||
|
service.getWorkspaceRoot(getTest1Project()),
|
||||||
|
'branch-b.md',
|
||||||
|
);
|
||||||
|
const text = await readFile(filePath, { encoding: 'utf-8' });
|
||||||
|
expect(text).toMatch(/Commit branch b/gi);
|
||||||
|
}, 30_000);
|
||||||
|
it('nonexistent branch', async () => {
|
||||||
|
return expect(
|
||||||
|
service.checkoutBranch(getTest1Project(), 'nonexistent'),
|
||||||
|
).rejects.toBeInstanceOf(NotFoundException);
|
||||||
|
}, 30_000);
|
||||||
|
it('checkout the specified version', async () => {
|
||||||
|
await service.checkoutBranch(getTest1Project(), 'master');
|
||||||
|
const filePath = join(
|
||||||
|
service.getWorkspaceRoot(getTest1Project()),
|
||||||
|
'README.md',
|
||||||
|
);
|
||||||
|
const text = await readFile(filePath, { encoding: 'utf-8' });
|
||||||
|
expect(text).toMatch(/Commit 1/gi);
|
||||||
|
}, 30_000);
|
||||||
});
|
});
|
||||||
|
|
||||||
describe.skip('checkout', () => {
|
describe('checkoutCommit', () => {
|
||||||
let task: PipelineTask;
|
|
||||||
let workspaceRoot: string;
|
|
||||||
beforeEach(() => {
|
|
||||||
const project = new Project();
|
|
||||||
const pipeline = new Pipeline();
|
|
||||||
task = new PipelineTask();
|
|
||||||
pipeline.project = project;
|
|
||||||
task.pipeline = pipeline;
|
|
||||||
project.id = 'pid';
|
|
||||||
project.name = 'pname';
|
|
||||||
pipeline.id = 'lid';
|
|
||||||
pipeline.name = 'pipeline';
|
|
||||||
task.id = 'tid';
|
|
||||||
task.commit = '123123hash';
|
|
||||||
workspaceRoot = service.getWorkspaceRootByTask(task);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be checkout', async () => {
|
it('should be checkout', async () => {
|
||||||
task.commit = '498c782685';
|
await service.checkoutCommit(getTest1Project(), '498c782685');
|
||||||
await service.checkout(task, workspaceRoot);
|
const filePath = join(
|
||||||
const filePath = join(workspaceRoot, 'README.md');
|
service.getWorkspaceRoot(getTest1Project()),
|
||||||
|
'README.md',
|
||||||
|
);
|
||||||
const text = await readFile(filePath, { encoding: 'utf-8' });
|
const text = await readFile(filePath, { encoding: 'utf-8' });
|
||||||
expect(text).toMatch(/Commit 1/gi);
|
expect(text).toMatch(/Commit 1/gi);
|
||||||
}, 20_000);
|
}, 20_000);
|
||||||
it('should be checkout right commit', async () => {
|
it('should be checkout right commit', async () => {
|
||||||
task.commit = '7f7123fe5b';
|
await service.checkoutCommit(getTest1Project(), '7f7123fe5b');
|
||||||
await service.checkout(task, workspaceRoot);
|
const filePath = join(
|
||||||
const filePath = join(workspaceRoot, 'README.md');
|
service.getWorkspaceRoot(getTest1Project()),
|
||||||
|
'README.md',
|
||||||
|
);
|
||||||
const text = await readFile(filePath, { encoding: 'utf-8' });
|
const text = await readFile(filePath, { encoding: 'utf-8' });
|
||||||
expect(text).toMatch(/(?!Commit 1)/gi);
|
expect(text).toMatch(/(?!Commit 1)/gi);
|
||||||
}, 20_000);
|
}, 20_000);
|
||||||
it('should be checkout right commit (复用)', async () => {
|
|
||||||
task.commit = '498c782685';
|
|
||||||
await service.checkout(task, workspaceRoot);
|
|
||||||
task.commit = '7f7123fe5b';
|
|
||||||
await service.checkout(task, workspaceRoot);
|
|
||||||
const filePath = join(workspaceRoot, 'README.md');
|
|
||||||
const text = await readFile(filePath, { encoding: 'utf-8' });
|
|
||||||
expect(text).toMatch(/(?!Commit 1)/gi);
|
|
||||||
}, 30_000);
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('getWorkspaceRootByTask', () => {
|
|
||||||
it('should be return right path', () => {
|
|
||||||
const project = new Project();
|
|
||||||
const pipeline = new Pipeline();
|
|
||||||
const task = new PipelineTask();
|
|
||||||
pipeline.project = project;
|
|
||||||
task.pipeline = pipeline;
|
|
||||||
project.id = 'pid';
|
|
||||||
project.name = 'pname';
|
|
||||||
pipeline.id = 'lid';
|
|
||||||
pipeline.name = 'pipeline/\\-名称';
|
|
||||||
task.id = 'tid';
|
|
||||||
task.commit = '123123hash';
|
|
||||||
|
|
||||||
expect(service.getWorkspaceRootByTask(task)).toMatch(
|
|
||||||
/\/pname\/pipeline%2F%5C-%E5%90%8D%E7%A7%B0-123123hash$/,
|
|
||||||
);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('fetch', () => {
|
|
||||||
it('success', async () => {
|
|
||||||
const project = new Project();
|
|
||||||
const pipeline = new Pipeline();
|
|
||||||
pipeline.branch = 'test';
|
|
||||||
const fetch = jest.fn((_: any) => Promise.resolve());
|
|
||||||
pipeline.project = project;
|
|
||||||
const getGit = jest.spyOn(service, 'getGit').mockImplementation(() =>
|
|
||||||
Promise.resolve({
|
|
||||||
fetch,
|
|
||||||
} as any),
|
|
||||||
);
|
|
||||||
await expect(service.fetch(pipeline)).resolves.toEqual(getInstanceName());
|
|
||||||
expect(getGit).toBeCalledTimes(1);
|
|
||||||
expect(getGit.mock.calls[0]?.[0]).toEqual(project);
|
|
||||||
expect(fetch).toBeCalledTimes(1);
|
|
||||||
expect(fetch.mock.calls[0]?.[0]).toMatchObject([
|
|
||||||
'origin',
|
|
||||||
'test',
|
|
||||||
'--depth=100',
|
|
||||||
]);
|
|
||||||
});
|
|
||||||
it('failed a', async () => {
|
|
||||||
const project = new Project();
|
|
||||||
const pipeline = new Pipeline();
|
|
||||||
pipeline.branch = 'test';
|
|
||||||
const fetch = jest.fn((_: any) => Promise.resolve());
|
|
||||||
pipeline.project = project;
|
|
||||||
const getGit = jest
|
|
||||||
.spyOn(service, 'getGit')
|
|
||||||
.mockImplementation(() => Promise.reject('error'));
|
|
||||||
await expect(service.fetch(pipeline)).resolves.toMatchObject(new Nack());
|
|
||||||
expect(getGit).toBeCalledTimes(1);
|
|
||||||
expect(getGit.mock.calls[0]?.[0]).toEqual(project);
|
|
||||||
expect(fetch).toBeCalledTimes(0);
|
|
||||||
});
|
|
||||||
it('failed b', async () => {
|
|
||||||
const project = new Project();
|
|
||||||
const pipeline = new Pipeline();
|
|
||||||
pipeline.branch = 'test';
|
|
||||||
const fetch = jest.fn((_: any) => Promise.reject('error'));
|
|
||||||
pipeline.project = project;
|
|
||||||
const getGit = jest.spyOn(service, 'getGit').mockImplementation(() =>
|
|
||||||
Promise.resolve({
|
|
||||||
fetch,
|
|
||||||
} as any),
|
|
||||||
);
|
|
||||||
await expect(service.fetch(pipeline)).resolves.toMatchObject(new Nack());
|
|
||||||
expect(getGit).toBeCalledTimes(1);
|
|
||||||
expect(fetch).toBeCalledTimes(1);
|
|
||||||
});
|
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
@ -1,5 +1,3 @@
|
|||||||
import { ListLogsOption } from './models/list-logs.options';
|
|
||||||
import { PipelineTask } from './../pipeline-tasks/pipeline-task.entity';
|
|
||||||
import { Injectable, NotFoundException } from '@nestjs/common';
|
import { Injectable, NotFoundException } from '@nestjs/common';
|
||||||
import { InjectRepository } from '@nestjs/typeorm';
|
import { InjectRepository } from '@nestjs/typeorm';
|
||||||
import { F_OK } from 'constants';
|
import { F_OK } from 'constants';
|
||||||
@ -9,89 +7,57 @@ import { gitP } from 'simple-git';
|
|||||||
import { Repository } from 'typeorm';
|
import { Repository } from 'typeorm';
|
||||||
import { Project } from '../projects/project.entity';
|
import { Project } from '../projects/project.entity';
|
||||||
import { ListBranchesArgs } from './dtos/list-branches.args';
|
import { ListBranchesArgs } from './dtos/list-branches.args';
|
||||||
|
import { ListLogsArgs } from './dtos/list-logs.args';
|
||||||
import { ConfigService } from '@nestjs/config';
|
import { ConfigService } from '@nestjs/config';
|
||||||
import { Commit } from './dtos/log-list.model';
|
|
||||||
import { Nack, RabbitRPC, RabbitSubscribe } from '@golevelup/nestjs-rabbitmq';
|
|
||||||
import { Pipeline } from '../pipelines/pipeline.entity';
|
|
||||||
import { InjectPinoLogger, PinoLogger } from 'nestjs-pino';
|
|
||||||
import {
|
|
||||||
EXCHANGE_REPO,
|
|
||||||
QUEUE_FETCH,
|
|
||||||
QUEUE_LIST_COMMITS,
|
|
||||||
QUEUE_REFRESH_REPO,
|
|
||||||
ROUTE_FETCH,
|
|
||||||
ROUTE_LIST_COMMITS,
|
|
||||||
} from './repos.constants';
|
|
||||||
import { getSelfInstanceQueueKey } from '../commons/utils/rabbit-mq';
|
|
||||||
import {
|
|
||||||
getInstanceName,
|
|
||||||
getSelfInstanceRouteKey,
|
|
||||||
} from '../commons/utils/rabbit-mq';
|
|
||||||
import { ApplicationException } from '../commons/exceptions/application.exception';
|
|
||||||
import {
|
|
||||||
EXCHANGE_PROJECT_FANOUT,
|
|
||||||
ROUTE_PROJECT_CHANGE,
|
|
||||||
} from '../projects/projects.constants';
|
|
||||||
import { RedisMutexService } from '../commons/redis-mutex/redis-mutex.service';
|
|
||||||
import { rm } from 'fs/promises';
|
|
||||||
|
|
||||||
const DEFAULT_REMOTE_NAME = 'origin';
|
const DEFAULT_REMOTE_NAME = 'origin';
|
||||||
const INFO_PATH = '@info';
|
|
||||||
@Injectable()
|
@Injectable()
|
||||||
export class ReposService {
|
export class ReposService {
|
||||||
constructor(
|
constructor(
|
||||||
@InjectRepository(Project)
|
@InjectRepository(Project)
|
||||||
private readonly projectRepository: Repository<Project>,
|
private readonly projectRepository: Repository<Project>,
|
||||||
private readonly configService: ConfigService,
|
private readonly configService: ConfigService,
|
||||||
@InjectPinoLogger(ReposService.name)
|
|
||||||
private readonly logger: PinoLogger,
|
|
||||||
private readonly redisMutexService: RedisMutexService,
|
|
||||||
) {}
|
) {}
|
||||||
|
|
||||||
getWorkspaceRoot(project: Project): string {
|
getWorkspaceRoot(project: Project): string {
|
||||||
return join(
|
return join(
|
||||||
this.configService.get<string>('workspaces.root'),
|
this.configService.get<string>('workspaces.root'),
|
||||||
encodeURIComponent(project.name),
|
project.name,
|
||||||
INFO_PATH,
|
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
getDeployRoot(task: PipelineTask) {
|
async lockWorkspace(workspaceRoot: string) {
|
||||||
return join(
|
// TODO: 获取锁,失败抛错。
|
||||||
this.configService.get<string>('workspaces.root'),
|
|
||||||
encodeURIComponent(task.pipeline.project.name),
|
|
||||||
encodeURIComponent(`deploy-${task.pipeline.name}`),
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async getGit(
|
async getGit(project: Project) {
|
||||||
project: Project,
|
const workspaceRoot = this.getWorkspaceRoot(project);
|
||||||
workspaceRoot?: string,
|
await this.lockWorkspace(workspaceRoot);
|
||||||
{ fetch = true } = {},
|
|
||||||
) {
|
|
||||||
if (!workspaceRoot) {
|
|
||||||
workspaceRoot = this.getWorkspaceRoot(project);
|
|
||||||
}
|
|
||||||
|
|
||||||
await access(workspaceRoot, F_OK).catch(async () => {
|
const firstInit = await access(workspaceRoot, F_OK)
|
||||||
await mkdir(workspaceRoot, { recursive: true });
|
.then(() => false)
|
||||||
});
|
.catch(async () => {
|
||||||
|
await mkdir(workspaceRoot);
|
||||||
|
return true;
|
||||||
|
});
|
||||||
const git = gitP(workspaceRoot);
|
const git = gitP(workspaceRoot);
|
||||||
if (!(await git.checkIsRepo().catch(() => false))) {
|
if (firstInit) {
|
||||||
await git.init();
|
await git.init();
|
||||||
await git.addRemote(DEFAULT_REMOTE_NAME, project.sshUrl);
|
await git.addRemote(DEFAULT_REMOTE_NAME, project.sshUrl);
|
||||||
}
|
}
|
||||||
if (fetch) {
|
|
||||||
await git.fetch();
|
|
||||||
}
|
|
||||||
return git;
|
return git;
|
||||||
}
|
}
|
||||||
|
|
||||||
async listLogs({ project, branch }: ListLogsOption) {
|
async listLogs(dto: ListLogsArgs) {
|
||||||
|
const project = await this.projectRepository.findOneOrFail({
|
||||||
|
id: dto.projectId,
|
||||||
|
});
|
||||||
const git = await this.getGit(project);
|
const git = await this.getGit(project);
|
||||||
return await git.log(
|
await git.fetch();
|
||||||
branch ? ['--branches', `remotes/origin/${branch}`, '--'] : ['--all'],
|
return await git.log({
|
||||||
);
|
'--branches': dto.branch ?? '',
|
||||||
|
'--remotes': DEFAULT_REMOTE_NAME,
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
async listBranches(dto: ListBranchesArgs) {
|
async listBranches(dto: ListBranchesArgs) {
|
||||||
@ -102,8 +68,26 @@ export class ReposService {
|
|||||||
return git.branch();
|
return git.branch();
|
||||||
}
|
}
|
||||||
|
|
||||||
async checkout(task: PipelineTask, workspaceRoot: string) {
|
async checkoutBranch(project: Project, branch: string) {
|
||||||
const git = await this.getGit(task.pipeline.project, workspaceRoot);
|
const git = await this.getGit(project);
|
||||||
|
try {
|
||||||
|
await git.fetch(DEFAULT_REMOTE_NAME, branch);
|
||||||
|
} catch (err) {
|
||||||
|
if (err.message.includes("couldn't find remote ref nonexistent")) {
|
||||||
|
throw new NotFoundException(err.message);
|
||||||
|
}
|
||||||
|
throw err;
|
||||||
|
}
|
||||||
|
await git.checkout([
|
||||||
|
'-B',
|
||||||
|
branch,
|
||||||
|
'--track',
|
||||||
|
`${DEFAULT_REMOTE_NAME}/${branch}`,
|
||||||
|
]);
|
||||||
|
}
|
||||||
|
|
||||||
|
async checkoutCommit(project: Project, commitNumber: string) {
|
||||||
|
const git = await this.getGit(project);
|
||||||
try {
|
try {
|
||||||
await git.fetch(DEFAULT_REMOTE_NAME);
|
await git.fetch(DEFAULT_REMOTE_NAME);
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
@ -112,129 +96,6 @@ export class ReposService {
|
|||||||
}
|
}
|
||||||
throw err;
|
throw err;
|
||||||
}
|
}
|
||||||
await git.checkout([task.commit]);
|
await git.checkout([commitNumber]);
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* get workspace root absolute path
|
|
||||||
*
|
|
||||||
* ! example: `/var/tmp/fennec-workspaces-root/project/pipeline_name-commit_hash`
|
|
||||||
* @param task {PipelineTask} task (with pipeline and project info)
|
|
||||||
*/
|
|
||||||
getWorkspaceRootByTask(task: PipelineTask) {
|
|
||||||
return join(
|
|
||||||
this.configService.get<string>('workspaces.root'),
|
|
||||||
encodeURIComponent(task.pipeline.project.name),
|
|
||||||
encodeURIComponent(`${task.pipeline.name}-${task.commit}`),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
async checkout4Task(task: PipelineTask): Promise<string> {
|
|
||||||
const path = this.getWorkspaceRootByTask(task);
|
|
||||||
await this.checkout(task, path);
|
|
||||||
return path;
|
|
||||||
}
|
|
||||||
|
|
||||||
@RabbitRPC({
|
|
||||||
exchange: EXCHANGE_REPO,
|
|
||||||
routingKey: [
|
|
||||||
ROUTE_LIST_COMMITS,
|
|
||||||
getSelfInstanceRouteKey(ROUTE_LIST_COMMITS),
|
|
||||||
],
|
|
||||||
queue: getSelfInstanceQueueKey(QUEUE_LIST_COMMITS),
|
|
||||||
queueOptions: {
|
|
||||||
autoDelete: true,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async listCommits(pipeline: Pipeline): Promise<[Error, Commit[]?]> {
|
|
||||||
const git = await this.getGit(pipeline.project, undefined, {
|
|
||||||
fetch: false,
|
|
||||||
});
|
|
||||||
try {
|
|
||||||
const data = await git.log([
|
|
||||||
'-100',
|
|
||||||
'--branches',
|
|
||||||
`remotes/origin/${pipeline.branch}`,
|
|
||||||
'--',
|
|
||||||
]);
|
|
||||||
return [
|
|
||||||
null,
|
|
||||||
data.all.map(
|
|
||||||
(it) =>
|
|
||||||
({
|
|
||||||
...it,
|
|
||||||
date: new Date(it.date),
|
|
||||||
} as Commit),
|
|
||||||
),
|
|
||||||
];
|
|
||||||
} catch (error) {
|
|
||||||
this.logger.error(
|
|
||||||
{ error, pipeline },
|
|
||||||
'[listCommits] %s',
|
|
||||||
error?.message,
|
|
||||||
);
|
|
||||||
return [new ApplicationException(error)];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
@RabbitRPC({
|
|
||||||
exchange: EXCHANGE_REPO,
|
|
||||||
routingKey: [ROUTE_FETCH, getSelfInstanceRouteKey(ROUTE_FETCH)],
|
|
||||||
queue: getSelfInstanceQueueKey(QUEUE_FETCH),
|
|
||||||
queueOptions: {
|
|
||||||
autoDelete: true,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async fetch(pipeline: Pipeline): Promise<string | null | Nack> {
|
|
||||||
const unlock = await this.redisMutexService.lock(
|
|
||||||
`repo-project-${pipeline.projectId}`,
|
|
||||||
);
|
|
||||||
try {
|
|
||||||
const git = await this.getGit(pipeline.project, undefined, {
|
|
||||||
fetch: false,
|
|
||||||
});
|
|
||||||
await git.fetch(['origin', pipeline.branch, '--depth=100']);
|
|
||||||
return getInstanceName();
|
|
||||||
} catch (error) {
|
|
||||||
this.logger.error({ error, pipeline }, '[fetch] %s', error?.message);
|
|
||||||
return new Nack();
|
|
||||||
} finally {
|
|
||||||
await unlock();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
@RabbitSubscribe({
|
|
||||||
exchange: EXCHANGE_PROJECT_FANOUT,
|
|
||||||
routingKey: ROUTE_PROJECT_CHANGE,
|
|
||||||
queue: QUEUE_REFRESH_REPO,
|
|
||||||
queueOptions: {
|
|
||||||
autoDelete: true,
|
|
||||||
durable: true,
|
|
||||||
},
|
|
||||||
})
|
|
||||||
async refreshRepo([project]: [Project]) {
|
|
||||||
this.logger.info({ project }, '[refreshRepo] start');
|
|
||||||
const unlock = await this.redisMutexService.lock(
|
|
||||||
`repo-project-${project.id}`,
|
|
||||||
{
|
|
||||||
timeout: null,
|
|
||||||
},
|
|
||||||
);
|
|
||||||
try {
|
|
||||||
const path = join(
|
|
||||||
this.configService.get<string>('workspaces.root'),
|
|
||||||
encodeURIComponent(project.name),
|
|
||||||
);
|
|
||||||
await rm(path, { recursive: true });
|
|
||||||
this.logger.info({ project }, '[refreshRepo] success');
|
|
||||||
} catch (error) {
|
|
||||||
this.logger.error(
|
|
||||||
{ project, error },
|
|
||||||
'[refreshRepo] failed. $s',
|
|
||||||
error.message,
|
|
||||||
);
|
|
||||||
} finally {
|
|
||||||
await unlock();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
8
src/repos/workspace-action.consumer.ts
Normal file
8
src/repos/workspace-action.consumer.ts
Normal file
@ -0,0 +1,8 @@
|
|||||||
|
import { Process, Processor } from '@nestjs/bull';
|
||||||
|
import { WORKSPACE_ACTION } from './repos.constants';
|
||||||
|
|
||||||
|
@Processor(WORKSPACE_ACTION)
|
||||||
|
export class WorkspaceActionConsumer {
|
||||||
|
@Process()
|
||||||
|
async dispatch() {}
|
||||||
|
}
|
5
src/repos/workspace-action.model.ts
Normal file
5
src/repos/workspace-action.model.ts
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
import { WorkspaceActions } from './workspace-actions.enum';
|
||||||
|
|
||||||
|
export class WorkspaceAction {
|
||||||
|
action: WorkspaceActions;
|
||||||
|
}
|
6
src/repos/workspace-actions.enum.ts
Normal file
6
src/repos/workspace-actions.enum.ts
Normal file
@ -0,0 +1,6 @@
|
|||||||
|
export enum WorkspaceActions {
|
||||||
|
checkoutBranch = 'checkoutBranch',
|
||||||
|
checkoutCommit = 'checkoutCommit',
|
||||||
|
listLogs = 'listLogs',
|
||||||
|
listBranches = 'listBranches',
|
||||||
|
}
|
@ -1,8 +0,0 @@
|
|||||||
import { IsString } from 'class-validator';
|
|
||||||
|
|
||||||
export class GiteaHookPayloadDto {
|
|
||||||
@IsString()
|
|
||||||
ref: string;
|
|
||||||
@IsString()
|
|
||||||
after: string;
|
|
||||||
}
|
|
@ -1,3 +0,0 @@
|
|||||||
export enum SourceService {
|
|
||||||
gitea = 'gitea',
|
|
||||||
}
|
|
@ -1,25 +0,0 @@
|
|||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { GiteaWebhooksController } from './gitea-webhooks.controller';
|
|
||||||
import { WebhooksService } from './webhooks.service';
|
|
||||||
|
|
||||||
describe('GiteaWebhooksController', () => {
|
|
||||||
let controller: GiteaWebhooksController;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
controllers: [GiteaWebhooksController],
|
|
||||||
providers: [
|
|
||||||
{
|
|
||||||
provide: WebhooksService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
controller = module.get<GiteaWebhooksController>(GiteaWebhooksController);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(controller).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,37 +0,0 @@
|
|||||||
import { Body, Controller, Headers, Param, Post } from '@nestjs/common';
|
|
||||||
import { validateOrReject } from 'class-validator';
|
|
||||||
import { pick } from 'ramda';
|
|
||||||
import { GiteaHookPayloadDto } from './dtos/gitea-hook-payload.dto';
|
|
||||||
import { SourceService } from './enums/source-service.enum';
|
|
||||||
import { WebhookLog } from './webhook-log.entity';
|
|
||||||
import { WebhooksService } from './webhooks.service';
|
|
||||||
|
|
||||||
@Controller('gitea-webhooks')
|
|
||||||
export class GiteaWebhooksController {
|
|
||||||
constructor(private readonly service: WebhooksService) {}
|
|
||||||
@Post(':pipelineId')
|
|
||||||
async onCall(
|
|
||||||
@Headers('X-Gitea-Delivery') delivery: string,
|
|
||||||
@Headers('X-Gitea-Event') event: string,
|
|
||||||
@Headers('X-Gitea-Signature') signature: string,
|
|
||||||
@Body() body: Buffer,
|
|
||||||
@Param('pipelineId') pipelineId: string,
|
|
||||||
) {
|
|
||||||
const payload = Object.assign(
|
|
||||||
new GiteaHookPayloadDto(),
|
|
||||||
JSON.parse(body.toString('utf-8')),
|
|
||||||
);
|
|
||||||
await validateOrReject(payload);
|
|
||||||
await this.service.verifySignature(body, signature, 'boardcat');
|
|
||||||
return await this.service
|
|
||||||
.onCall(pipelineId, {
|
|
||||||
payload,
|
|
||||||
sourceDelivery: delivery,
|
|
||||||
sourceEvent: event,
|
|
||||||
sourceService: SourceService.gitea,
|
|
||||||
})
|
|
||||||
.then((data) =>
|
|
||||||
pick<keyof WebhookLog>(['id', 'createdAt', 'localEvent'])(data),
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,9 +0,0 @@
|
|||||||
import { SourceService } from '../enums/source-service.enum';
|
|
||||||
import { WebhookLog } from '../webhook-log.entity';
|
|
||||||
|
|
||||||
export class CreateWebhookLogModel<T> implements Partial<WebhookLog> {
|
|
||||||
sourceDelivery: string;
|
|
||||||
sourceEvent: string;
|
|
||||||
sourceService: SourceService;
|
|
||||||
payload: T;
|
|
||||||
}
|
|
@ -1,19 +0,0 @@
|
|||||||
import { Column, Entity } from 'typeorm';
|
|
||||||
import { AppBaseEntity } from './../commons/entities/app-base-entity';
|
|
||||||
import { SourceService } from './enums/source-service.enum';
|
|
||||||
|
|
||||||
@Entity()
|
|
||||||
export class WebhookLog extends AppBaseEntity {
|
|
||||||
@Column()
|
|
||||||
sourceDelivery: string;
|
|
||||||
@Column({ type: 'enum', enum: SourceService })
|
|
||||||
sourceService: SourceService;
|
|
||||||
@Column()
|
|
||||||
sourceEvent: string;
|
|
||||||
@Column({ type: 'jsonb' })
|
|
||||||
payload: any;
|
|
||||||
@Column()
|
|
||||||
localEvent: string;
|
|
||||||
@Column({ type: 'jsonb' })
|
|
||||||
localPayload: any;
|
|
||||||
}
|
|
@ -1,14 +0,0 @@
|
|||||||
import { MiddlewareConsumer, Module } from '@nestjs/common';
|
|
||||||
import { TypeOrmModule } from '@nestjs/typeorm';
|
|
||||||
import { PipelineTasksModule } from '../pipeline-tasks/pipeline-tasks.module';
|
|
||||||
import { GiteaWebhooksController } from './gitea-webhooks.controller';
|
|
||||||
import { WebhookLog } from './webhook-log.entity';
|
|
||||||
import { WebhooksService } from './webhooks.service';
|
|
||||||
|
|
||||||
@Module({
|
|
||||||
imports: [TypeOrmModule.forFeature([WebhookLog]), PipelineTasksModule],
|
|
||||||
controllers: [GiteaWebhooksController],
|
|
||||||
providers: [WebhooksService],
|
|
||||||
})
|
|
||||||
export class WebhooksModule {
|
|
||||||
}
|
|
@ -1,57 +0,0 @@
|
|||||||
import { UnauthorizedException } from '@nestjs/common';
|
|
||||||
import { Test, TestingModule } from '@nestjs/testing';
|
|
||||||
import { getRepositoryToken } from '@nestjs/typeorm';
|
|
||||||
import { readFile } from 'fs/promises';
|
|
||||||
import { join } from 'path';
|
|
||||||
import { Repository } from 'typeorm';
|
|
||||||
import { PipelineTasksService } from '../pipeline-tasks/pipeline-tasks.service';
|
|
||||||
import { WebhookLog } from './webhook-log.entity';
|
|
||||||
import { WebhooksService } from './webhooks.service';
|
|
||||||
|
|
||||||
describe('WebhooksService', () => {
|
|
||||||
let service: WebhooksService;
|
|
||||||
|
|
||||||
beforeEach(async () => {
|
|
||||||
const module: TestingModule = await Test.createTestingModule({
|
|
||||||
providers: [
|
|
||||||
WebhooksService,
|
|
||||||
{
|
|
||||||
provide: PipelineTasksService,
|
|
||||||
useValue: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
provide: getRepositoryToken(WebhookLog),
|
|
||||||
useValue: new Repository(),
|
|
||||||
},
|
|
||||||
],
|
|
||||||
}).compile();
|
|
||||||
|
|
||||||
service = module.get<WebhooksService>(WebhooksService);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should be defined', () => {
|
|
||||||
expect(service).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('verifySignature', () => {
|
|
||||||
const signature =
|
|
||||||
'b175e07189a6106f386b62253b18b5879c4b1f3af2f11fe13a294602671e361a';
|
|
||||||
const secret = 'boardcat';
|
|
||||||
let payload: Buffer;
|
|
||||||
beforeAll(async () => {
|
|
||||||
payload = await readFile(
|
|
||||||
join(__dirname, '../../test/data/gitea-hook-payload.json.bin'),
|
|
||||||
);
|
|
||||||
});
|
|
||||||
it('must be valid', async () => {
|
|
||||||
await expect(
|
|
||||||
service.verifySignature(payload, signature, secret),
|
|
||||||
).resolves.toEqual(undefined);
|
|
||||||
});
|
|
||||||
it('must be invalid', async () => {
|
|
||||||
await expect(
|
|
||||||
service.verifySignature(payload, 'test', secret),
|
|
||||||
).rejects.toThrowError(UnauthorizedException);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
@ -1,65 +0,0 @@
|
|||||||
import {
|
|
||||||
BadRequestException,
|
|
||||||
Injectable,
|
|
||||||
UnauthorizedException,
|
|
||||||
} from '@nestjs/common';
|
|
||||||
import { InjectRepository } from '@nestjs/typeorm';
|
|
||||||
import { createHmac } from 'crypto';
|
|
||||||
import { Repository } from 'typeorm';
|
|
||||||
import { PipelineUnits } from '../pipeline-tasks/enums/pipeline-units.enum';
|
|
||||||
import { PipelineTasksService } from '../pipeline-tasks/pipeline-tasks.service';
|
|
||||||
import { GiteaHookPayloadDto } from './dtos/gitea-hook-payload.dto';
|
|
||||||
import { CreateWebhookLogModel } from './models/create-webhook-log.model';
|
|
||||||
import { WebhookLog } from './webhook-log.entity';
|
|
||||||
|
|
||||||
@Injectable()
|
|
||||||
export class WebhooksService {
|
|
||||||
constructor(
|
|
||||||
@InjectRepository(WebhookLog)
|
|
||||||
private readonly repository: Repository<WebhookLog>,
|
|
||||||
private readonly taskService: PipelineTasksService,
|
|
||||||
) {}
|
|
||||||
|
|
||||||
async onCall(
|
|
||||||
pipelineId: string,
|
|
||||||
model: CreateWebhookLogModel<GiteaHookPayloadDto>,
|
|
||||||
) {
|
|
||||||
if (model.sourceEvent.toLowerCase() === 'push') {
|
|
||||||
const taskDto = {
|
|
||||||
pipelineId,
|
|
||||||
commit: model.payload.after,
|
|
||||||
units: Object.values(PipelineUnits),
|
|
||||||
};
|
|
||||||
await this.taskService.addTask(taskDto);
|
|
||||||
return await this.repository.save(
|
|
||||||
this.repository.create({
|
|
||||||
...model,
|
|
||||||
localEvent: 'create-pipeline-task',
|
|
||||||
localPayload: taskDto,
|
|
||||||
}),
|
|
||||||
);
|
|
||||||
} else {
|
|
||||||
throw new BadRequestException('无法处理的请求');
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
async verifySignature(payload: any, signature: string, secret: string) {
|
|
||||||
const local = await new Promise<string>((resolve, reject) => {
|
|
||||||
const hmac = createHmac('sha256', secret);
|
|
||||||
hmac.on('readable', () => {
|
|
||||||
const data = hmac.read();
|
|
||||||
if (data) {
|
|
||||||
resolve(data.toString('hex'));
|
|
||||||
}
|
|
||||||
});
|
|
||||||
hmac.on('error', (err) => {
|
|
||||||
reject(err);
|
|
||||||
});
|
|
||||||
hmac.write(payload);
|
|
||||||
hmac.end();
|
|
||||||
});
|
|
||||||
if (local !== signature) {
|
|
||||||
throw new UnauthorizedException();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,15 +0,0 @@
|
|||||||
module.exports = {
|
|
||||||
apps: [
|
|
||||||
{
|
|
||||||
name: 'app',
|
|
||||||
script: __dirname + '/index.js',
|
|
||||||
watch: false,
|
|
||||||
ignore_watch: ['node_modules'],
|
|
||||||
log_date_format: 'MM-DD HH:mm:ss.SSS Z',
|
|
||||||
env: {},
|
|
||||||
max_restarts: 5,
|
|
||||||
kill_timeout: 10_000,
|
|
||||||
wait_ready: true,
|
|
||||||
},
|
|
||||||
],
|
|
||||||
};
|
|
@ -1,23 +0,0 @@
|
|||||||
import { createServer } from 'http';
|
|
||||||
|
|
||||||
var app = createServer(function (req, res) {
|
|
||||||
res.writeHead(200);
|
|
||||||
setTimeout(() => {
|
|
||||||
res.end('hey');
|
|
||||||
}, 2000);
|
|
||||||
});
|
|
||||||
|
|
||||||
var listener = app.listen(0, function () {
|
|
||||||
console.log('Listening on port ' + listener.address().port);
|
|
||||||
setTimeout(() => {
|
|
||||||
// Here we send the ready signal to PM2
|
|
||||||
process.send('ready');
|
|
||||||
}, 5000);
|
|
||||||
});
|
|
||||||
|
|
||||||
process.on('SIGINT', function () {
|
|
||||||
listener.close();
|
|
||||||
setTimeout(() => {
|
|
||||||
process.exit(0);
|
|
||||||
}, 2000);
|
|
||||||
});
|
|
@ -1,10 +0,0 @@
|
|||||||
{
|
|
||||||
"name": "deploy-service",
|
|
||||||
"version": "1.0.0",
|
|
||||||
"description": "For Test",
|
|
||||||
"main": "index.js",
|
|
||||||
"type": "module",
|
|
||||||
"scripts": {},
|
|
||||||
"author": "Ivan Li",
|
|
||||||
"license": "ISC"
|
|
||||||
}
|
|
@ -1,16 +0,0 @@
|
|||||||
import { createServer } from 'http';
|
|
||||||
|
|
||||||
var app = createServer(function (req, res) {
|
|
||||||
res.writeHead(200);
|
|
||||||
setTimeout(() => {
|
|
||||||
res.end('pm2');
|
|
||||||
}, 2000);
|
|
||||||
});
|
|
||||||
|
|
||||||
var listener = app.listen(33333, function () {
|
|
||||||
console.log('Listening on port ' + listener.address().port);
|
|
||||||
setTimeout(() => {
|
|
||||||
// Here we send the ready signal to PM2
|
|
||||||
process.send('ready');
|
|
||||||
}, 5000);
|
|
||||||
});
|
|
@ -1,8 +0,0 @@
|
|||||||
for (let i = 1; i <= 5; i++) {
|
|
||||||
console.log(i * 10);
|
|
||||||
}
|
|
||||||
console.error('Error Message');
|
|
||||||
console.error('Error Message 2');
|
|
||||||
console.log('Bye-bye');
|
|
||||||
|
|
||||||
process.exit(1);
|
|
@ -1,115 +0,0 @@
|
|||||||
{
|
|
||||||
"secret": "boardcat",
|
|
||||||
"ref": "refs/heads/master",
|
|
||||||
"before": "429de1eaedf1da83f1e0e3ac3d8b20e771b7051c",
|
|
||||||
"after": "429de1eaedf1da83f1e0e3ac3d8b20e771b7051c",
|
|
||||||
"compare_url": "",
|
|
||||||
"commits": [
|
|
||||||
{
|
|
||||||
"id": "429de1eaedf1da83f1e0e3ac3d8b20e771b7051c",
|
|
||||||
"message": "test(pipeline-tasks): pass test cases.\n",
|
|
||||||
"url": "https://git.ivanli.cc/Fennec/fennec-be/commit/429de1eaedf1da83f1e0e3ac3d8b20e771b7051c",
|
|
||||||
"author": {
|
|
||||||
"name": "Ivan",
|
|
||||||
"email": "ivanli@live.cn",
|
|
||||||
"username": ""
|
|
||||||
},
|
|
||||||
"committer": {
|
|
||||||
"name": "Ivan",
|
|
||||||
"email": "ivanli@live.cn",
|
|
||||||
"username": ""
|
|
||||||
},
|
|
||||||
"verification": null,
|
|
||||||
"timestamp": "0001-01-01T00:00:00Z",
|
|
||||||
"added": null,
|
|
||||||
"removed": null,
|
|
||||||
"modified": null
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"head_commit": null,
|
|
||||||
"repository": {
|
|
||||||
"id": 3,
|
|
||||||
"owner": {
|
|
||||||
"id": 3,
|
|
||||||
"login": "Fennec",
|
|
||||||
"full_name": "",
|
|
||||||
"email": "",
|
|
||||||
"avatar_url": "https://git.ivanli.cc/user/avatar/Fennec/-1",
|
|
||||||
"language": "",
|
|
||||||
"is_admin": false,
|
|
||||||
"last_login": "1970-01-01T08:00:00+08:00",
|
|
||||||
"created": "2021-01-30T16:46:11+08:00",
|
|
||||||
"username": "Fennec"
|
|
||||||
},
|
|
||||||
"name": "fennec-be",
|
|
||||||
"full_name": "Fennec/fennec-be",
|
|
||||||
"description": "Fennec CI/CD Back-End",
|
|
||||||
"empty": false,
|
|
||||||
"private": false,
|
|
||||||
"fork": false,
|
|
||||||
"template": false,
|
|
||||||
"parent": null,
|
|
||||||
"mirror": false,
|
|
||||||
"size": 1897,
|
|
||||||
"html_url": "https://git.ivanli.cc/Fennec/fennec-be",
|
|
||||||
"ssh_url": "ssh://gitea@git.ivanli.cc:7018/Fennec/fennec-be.git",
|
|
||||||
"clone_url": "https://git.ivanli.cc/Fennec/fennec-be.git",
|
|
||||||
"original_url": "",
|
|
||||||
"website": "",
|
|
||||||
"stars_count": 1,
|
|
||||||
"forks_count": 0,
|
|
||||||
"watchers_count": 1,
|
|
||||||
"open_issues_count": 0,
|
|
||||||
"open_pr_counter": 0,
|
|
||||||
"release_counter": 0,
|
|
||||||
"default_branch": "master",
|
|
||||||
"archived": false,
|
|
||||||
"created_at": "2021-01-31T09:58:38+08:00",
|
|
||||||
"updated_at": "2021-03-27T15:57:00+08:00",
|
|
||||||
"permissions": {
|
|
||||||
"admin": false,
|
|
||||||
"push": false,
|
|
||||||
"pull": false
|
|
||||||
},
|
|
||||||
"has_issues": true,
|
|
||||||
"internal_tracker": {
|
|
||||||
"enable_time_tracker": true,
|
|
||||||
"allow_only_contributors_to_track_time": true,
|
|
||||||
"enable_issue_dependencies": true
|
|
||||||
},
|
|
||||||
"has_wiki": true,
|
|
||||||
"has_pull_requests": true,
|
|
||||||
"has_projects": true,
|
|
||||||
"ignore_whitespace_conflicts": false,
|
|
||||||
"allow_merge_commits": true,
|
|
||||||
"allow_rebase": true,
|
|
||||||
"allow_rebase_explicit": true,
|
|
||||||
"allow_squash_merge": true,
|
|
||||||
"avatar_url": "",
|
|
||||||
"internal": false
|
|
||||||
},
|
|
||||||
"pusher": {
|
|
||||||
"id": 1,
|
|
||||||
"login": "Ivan",
|
|
||||||
"full_name": "Ivan Li",
|
|
||||||
"email": "ivan@noreply.%(DOMAIN)s",
|
|
||||||
"avatar_url": "https://git.ivanli.cc/user/avatar/Ivan/-1",
|
|
||||||
"language": "zh-CN",
|
|
||||||
"is_admin": true,
|
|
||||||
"last_login": "2021-03-26T22:28:05+08:00",
|
|
||||||
"created": "2021-01-23T18:15:30+08:00",
|
|
||||||
"username": "Ivan"
|
|
||||||
},
|
|
||||||
"sender": {
|
|
||||||
"id": 1,
|
|
||||||
"login": "Ivan",
|
|
||||||
"full_name": "Ivan Li",
|
|
||||||
"email": "ivan@noreply.%(DOMAIN)s",
|
|
||||||
"avatar_url": "https://git.ivanli.cc/user/avatar/Ivan/-1",
|
|
||||||
"language": "zh-CN",
|
|
||||||
"is_admin": true,
|
|
||||||
"last_login": "2021-03-26T22:28:05+08:00",
|
|
||||||
"created": "2021-01-23T18:15:30+08:00",
|
|
||||||
"username": "Ivan"
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,7 +0,0 @@
|
|||||||
let timer;
|
|
||||||
let count = 0;
|
|
||||||
setTimeout(() => clearInterval(timer), 1_000);
|
|
||||||
|
|
||||||
timer = setInterval(() => {
|
|
||||||
console.log(++count * 10);
|
|
||||||
}, 95);
|
|
@ -7,10 +7,9 @@
|
|||||||
"experimentalDecorators": true,
|
"experimentalDecorators": true,
|
||||||
"allowSyntheticDefaultImports": true,
|
"allowSyntheticDefaultImports": true,
|
||||||
"target": "es2017",
|
"target": "es2017",
|
||||||
"lib": ["ES2021"],
|
|
||||||
"sourceMap": true,
|
"sourceMap": true,
|
||||||
"outDir": "./dist",
|
"outDir": "./dist",
|
||||||
"rootDir": "./src",
|
|
||||||
"baseUrl": "./",
|
"baseUrl": "./",
|
||||||
},
|
"incremental": true
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user