PeerTube/server/lib/job-queue/job-queue.ts

494 lines
15 KiB
TypeScript
Raw Normal View History

2022-08-08 10:42:08 +02:00
import {
2022-08-08 15:48:17 +02:00
FlowJob,
FlowProducer,
2022-08-08 10:42:08 +02:00
Job,
JobsOptions,
Queue,
QueueEvents,
QueueEventsOptions,
QueueOptions,
QueueScheduler,
QueueSchedulerOptions,
Worker,
WorkerOptions
} from 'bullmq'
2020-12-14 12:00:35 +01:00
import { jobStates } from '@server/helpers/custom-validators/jobs'
import { CONFIG } from '@server/initializers/config'
2020-12-14 12:00:35 +01:00
import { processVideoRedundancy } from '@server/lib/job-queue/handlers/video-redundancy'
2022-08-08 15:48:17 +02:00
import { pick, timeoutPromise } from '@shared/core-utils'
2020-04-23 09:32:53 +02:00
import {
ActivitypubFollowPayload,
ActivitypubHttpBroadcastPayload,
ActivitypubHttpFetcherPayload,
ActivitypubHttpUnicastPayload,
ActorKeysPayload,
DeleteResumableUploadMetaFilePayload,
EmailPayload,
2022-08-08 15:48:17 +02:00
FederateVideoPayload,
2020-04-23 09:32:53 +02:00
JobState,
JobType,
ManageVideoTorrentPayload,
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
MoveObjectStoragePayload,
2022-08-08 15:48:17 +02:00
NotifyPayload,
RefreshPayload,
VideoFileImportPayload,
VideoImportPayload,
VideoLiveEndingPayload,
VideoRedundancyPayload,
2022-03-22 16:58:49 +01:00
VideoStudioEditionPayload,
VideoTranscodingPayload
2020-04-23 09:32:53 +02:00
} from '../../../shared/models'
import { logger } from '../../helpers/logger'
import { JOB_ATTEMPTS, JOB_COMPLETED_LIFETIME, JOB_CONCURRENCY, JOB_TTL, REPEAT_JOBS, WEBSERVER } from '../../initializers/constants'
2022-08-02 15:29:00 +02:00
import { Hooks } from '../plugins/hooks'
import { processActivityPubCleaner } from './handlers/activitypub-cleaner'
2020-12-14 12:00:35 +01:00
import { processActivityPubFollow } from './handlers/activitypub-follow'
2020-04-23 09:32:53 +02:00
import { processActivityPubHttpBroadcast } from './handlers/activitypub-http-broadcast'
import { processActivityPubHttpFetcher } from './handlers/activitypub-http-fetcher'
import { processActivityPubHttpUnicast } from './handlers/activitypub-http-unicast'
import { refreshAPObject } from './handlers/activitypub-refresher'
import { processActorKeys } from './handlers/actor-keys'
2020-12-14 12:00:35 +01:00
import { processEmail } from './handlers/email'
2022-08-08 15:48:17 +02:00
import { processFederateVideo } from './handlers/federate-video'
import { processManageVideoTorrent } from './handlers/manage-video-torrent'
import { onMoveToObjectStorageFailure, processMoveToObjectStorage } from './handlers/move-to-object-storage'
2022-08-08 15:48:17 +02:00
import { processNotify } from './handlers/notify'
import { processVideoFileImport } from './handlers/video-file-import'
2020-12-14 12:00:35 +01:00
import { processVideoImport } from './handlers/video-import'
import { processVideoLiveEnding } from './handlers/video-live-ending'
2022-03-22 16:58:49 +01:00
import { processVideoStudioEdition } from './handlers/video-studio-edition'
2020-12-14 12:00:35 +01:00
import { processVideoTranscoding } from './handlers/video-transcoding'
import { processVideosViewsStats } from './handlers/video-views-stats'
2022-08-08 15:48:17 +02:00
export type CreateJobArgument =
{ type: 'activitypub-http-broadcast', payload: ActivitypubHttpBroadcastPayload } |
2022-06-17 14:08:13 +02:00
{ type: 'activitypub-http-broadcast-parallel', payload: ActivitypubHttpBroadcastPayload } |
{ type: 'activitypub-http-unicast', payload: ActivitypubHttpUnicastPayload } |
{ type: 'activitypub-http-fetcher', payload: ActivitypubHttpFetcherPayload } |
{ type: 'activitypub-http-cleaner', payload: {} } |
{ type: 'activitypub-follow', payload: ActivitypubFollowPayload } |
{ type: 'video-file-import', payload: VideoFileImportPayload } |
{ type: 'video-transcoding', payload: VideoTranscodingPayload } |
{ type: 'email', payload: EmailPayload } |
2018-08-29 16:26:25 +02:00
{ type: 'video-import', payload: VideoImportPayload } |
{ type: 'activitypub-refresher', payload: RefreshPayload } |
{ type: 'videos-views-stats', payload: {} } |
{ type: 'video-live-ending', payload: VideoLiveEndingPayload } |
{ type: 'actor-keys', payload: ActorKeysPayload } |
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
{ type: 'video-redundancy', payload: VideoRedundancyPayload } |
{ type: 'delete-resumable-upload-meta-file', payload: DeleteResumableUploadMetaFilePayload } |
2022-03-22 16:58:49 +01:00
{ type: 'video-studio-edition', payload: VideoStudioEditionPayload } |
{ type: 'manage-video-torrent', payload: ManageVideoTorrentPayload } |
2022-08-08 15:48:17 +02:00
{ type: 'notify', payload: NotifyPayload } |
{ type: 'move-to-object-storage', payload: MoveObjectStoragePayload } |
{ type: 'federate-video', payload: FederateVideoPayload }
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
export type CreateJobOptions = {
delay?: number
priority?: number
}
2021-08-27 14:32:44 +02:00
const handlers: { [id in JobType]: (job: Job) => Promise<any> } = {
'activitypub-http-broadcast': processActivityPubHttpBroadcast,
2022-06-17 14:08:13 +02:00
'activitypub-http-broadcast-parallel': processActivityPubHttpBroadcast,
'activitypub-http-unicast': processActivityPubHttpUnicast,
'activitypub-http-fetcher': processActivityPubHttpFetcher,
'activitypub-cleaner': processActivityPubCleaner,
'activitypub-follow': processActivityPubFollow,
'video-file-import': processVideoFileImport,
'video-transcoding': processVideoTranscoding,
'email': processEmail,
2018-08-29 16:26:25 +02:00
'video-import': processVideoImport,
'videos-views-stats': processVideosViewsStats,
2020-01-10 10:11:28 +01:00
'activitypub-refresher': refreshAPObject,
'video-live-ending': processVideoLiveEnding,
'actor-keys': processActorKeys,
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
'video-redundancy': processVideoRedundancy,
2022-02-11 10:51:33 +01:00
'move-to-object-storage': processMoveToObjectStorage,
'manage-video-torrent': processManageVideoTorrent,
2022-08-08 15:48:17 +02:00
'notify': processNotify,
'video-studio-edition': processVideoStudioEdition,
'federate-video': processFederateVideo
}
const errorHandlers: { [id in JobType]?: (job: Job, err: any) => Promise<any> } = {
'move-to-object-storage': onMoveToObjectStorageFailure
}
2018-07-10 17:02:20 +02:00
const jobTypes: JobType[] = [
'activitypub-follow',
2018-05-09 09:08:22 +02:00
'activitypub-http-broadcast',
2022-06-17 14:08:13 +02:00
'activitypub-http-broadcast-parallel',
2018-05-09 09:08:22 +02:00
'activitypub-http-fetcher',
2018-07-10 17:02:20 +02:00
'activitypub-http-unicast',
'activitypub-cleaner',
2018-07-10 17:02:20 +02:00
'email',
'video-transcoding',
'video-file-import',
2018-08-29 16:26:25 +02:00
'video-import',
'videos-views-stats',
2020-01-10 10:11:28 +01:00
'activitypub-refresher',
'video-redundancy',
'actor-keys',
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
'video-live-ending',
2022-02-11 10:51:33 +01:00
'move-to-object-storage',
'manage-video-torrent',
2022-08-08 15:48:17 +02:00
'video-studio-edition',
'notify',
'federate-video'
2018-05-09 09:08:22 +02:00
]
2022-05-18 12:01:02 +02:00
const silentFailure = new Set<JobType>([ 'activitypub-http-unicast' ])
class JobQueue {
private static instance: JobQueue
2022-08-08 10:42:08 +02:00
private workers: { [id in JobType]?: Worker } = {}
2021-08-27 14:32:44 +02:00
private queues: { [id in JobType]?: Queue } = {}
2022-08-08 10:42:08 +02:00
private queueSchedulers: { [id in JobType]?: QueueScheduler } = {}
private queueEvents: { [id in JobType]?: QueueEvents } = {}
2022-08-08 15:48:17 +02:00
private flowProducer: FlowProducer
private initialized = false
private jobRedisPrefix: string
2020-01-31 16:56:52 +01:00
private constructor () {
}
init (produceOnly = false) {
// Already initialized
if (this.initialized === true) return
this.initialized = true
2019-04-11 11:33:44 +02:00
this.jobRedisPrefix = 'bull-' + WEBSERVER.HOST
2022-01-14 15:06:33 +01:00
2022-08-08 10:42:08 +02:00
for (const handlerName of (Object.keys(handlers) as JobType[])) {
this.buildWorker(handlerName, produceOnly)
this.buildQueue(handlerName)
this.buildQueueScheduler(handlerName, produceOnly)
this.buildQueueEvent(handlerName, produceOnly)
}
2022-08-08 15:48:17 +02:00
this.flowProducer = new FlowProducer({
connection: this.getRedisConnection(),
prefix: this.jobRedisPrefix
})
2022-08-08 10:42:08 +02:00
this.addRepeatableJobs()
}
private buildWorker (handlerName: JobType, produceOnly: boolean) {
const workerOptions: WorkerOptions = {
autorun: !produceOnly,
concurrency: this.getJobConcurrency(handlerName),
prefix: this.jobRedisPrefix,
2022-08-08 10:42:08 +02:00
connection: this.getRedisConnection()
2018-07-10 17:02:20 +02:00
}
2018-01-30 13:27:07 +01:00
2022-08-08 10:42:08 +02:00
const handler = function (job: Job) {
const timeout = JOB_TTL[handlerName]
const p = handlers[handlerName](job)
2022-08-08 10:42:08 +02:00
if (!timeout) return p
2022-08-08 10:42:08 +02:00
return timeoutPromise(p, timeout)
}
2022-08-08 10:42:08 +02:00
const processor = async (jobArg: Job<any>) => {
const job = await Hooks.wrapObject(jobArg, 'filter:job-queue.process.params', { type: handlerName })
2022-08-02 15:29:00 +02:00
2022-08-08 10:42:08 +02:00
return Hooks.wrapPromiseFun(handler, job, 'filter:job-queue.process.result')
}
2018-08-03 09:27:30 +02:00
2022-08-08 10:42:08 +02:00
const worker = new Worker(handlerName, processor, workerOptions)
2022-05-18 12:01:02 +02:00
2022-08-08 10:42:08 +02:00
worker.on('failed', (job, err) => {
const logLevel = silentFailure.has(handlerName)
? 'debug'
: 'error'
2022-08-08 10:42:08 +02:00
logger.log(logLevel, 'Cannot execute job %s in queue %s.', job.id, handlerName, { payload: job.data, err })
2018-02-12 11:25:09 +01:00
2022-08-08 10:42:08 +02:00
if (errorHandlers[job.name]) {
errorHandlers[job.name](job, err)
.catch(err => logger.error('Cannot run error handler for job failure %d in queue %s.', job.id, handlerName, { err }))
}
})
2018-07-10 17:02:20 +02:00
2022-08-08 10:42:08 +02:00
worker.on('error', err => {
logger.error('Error in job queue %s.', handlerName, { err })
})
this.workers[handlerName] = worker
}
private buildQueue (handlerName: JobType) {
const queueOptions: QueueOptions = {
connection: this.getRedisConnection(),
prefix: this.jobRedisPrefix
}
2018-08-29 16:26:25 +02:00
2022-08-08 10:42:08 +02:00
this.queues[handlerName] = new Queue(handlerName, queueOptions)
}
private buildQueueScheduler (handlerName: JobType, produceOnly: boolean) {
const queueSchedulerOptions: QueueSchedulerOptions = {
autorun: !produceOnly,
connection: this.getRedisConnection(),
prefix: this.jobRedisPrefix,
maxStalledCount: 10
}
this.queueSchedulers[handlerName] = new QueueScheduler(handlerName, queueSchedulerOptions)
}
2022-08-08 10:42:08 +02:00
private buildQueueEvent (handlerName: JobType, produceOnly: boolean) {
const queueEventsOptions: QueueEventsOptions = {
autorun: !produceOnly,
connection: this.getRedisConnection(),
prefix: this.jobRedisPrefix
2018-07-30 18:49:54 +02:00
}
2022-08-08 10:42:08 +02:00
this.queueEvents[handlerName] = new QueueEvents(handlerName, queueEventsOptions)
}
private getRedisConnection () {
return {
password: CONFIG.REDIS.AUTH,
db: CONFIG.REDIS.DB,
host: CONFIG.REDIS.HOSTNAME,
port: CONFIG.REDIS.PORT,
path: CONFIG.REDIS.SOCKET
}
}
2022-08-08 15:48:17 +02:00
// ---------------------------------------------------------------------------
2022-08-08 10:42:08 +02:00
async terminate () {
const promises = Object.keys(this.workers)
.map(handlerName => {
const worker: Worker = this.workers[handlerName]
const queue: Queue = this.queues[handlerName]
const queueScheduler: QueueScheduler = this.queueSchedulers[handlerName]
const queueEvent: QueueEvents = this.queueEvents[handlerName]
return Promise.all([
worker.close(false),
queue.close(),
queueScheduler.close(),
queueEvent.close()
])
})
return Promise.all(promises)
2018-07-30 18:49:54 +02:00
}
async pause () {
2022-08-09 11:35:07 +02:00
for (const handlerName of Object.keys(this.workers)) {
const worker: Worker = this.workers[handlerName]
const queue: Queue = this.queues[handlerName]
2022-08-08 10:42:08 +02:00
await worker.pause()
2022-08-09 11:35:07 +02:00
await queue.pause()
}
}
2022-08-09 11:35:07 +02:00
async resume () {
for (const handlerName of Object.keys(this.workers)) {
const worker: Worker = this.workers[handlerName]
const queue: Queue = this.queues[handlerName]
2022-08-08 10:42:08 +02:00
worker.resume()
2022-08-09 11:35:07 +02:00
await queue.resume()
}
}
2022-08-08 15:48:17 +02:00
// ---------------------------------------------------------------------------
createJobAsync (options: CreateJobArgument & CreateJobOptions): void {
this.createJob(options)
.catch(err => logger.error('Cannot create job.', { err, options }))
2020-01-31 16:56:52 +01:00
}
2022-08-08 15:48:17 +02:00
async createJob (options: CreateJobArgument & CreateJobOptions) {
const queue: Queue = this.queues[options.type]
2018-07-10 17:02:20 +02:00
if (queue === undefined) {
2022-08-08 15:48:17 +02:00
logger.error('Unknown queue %s: cannot create job.', options.type)
2020-01-31 16:56:52 +01:00
return
2018-07-10 17:02:20 +02:00
}
2022-08-08 15:48:17 +02:00
const jobOptions = this.buildJobOptions(options.type as JobType, pick(options, [ 'priority', 'delay' ]))
return queue.add('job', options.payload, jobOptions)
}
async createSequentialJobFlow (...jobs: ((CreateJobArgument & CreateJobOptions) | undefined)[]) {
let lastJob: FlowJob
for (const job of jobs) {
if (!job) continue
lastJob = {
2022-08-09 09:09:31 +02:00
...this.buildJobFlowOption(job),
2022-08-08 15:48:17 +02:00
children: lastJob
? [ lastJob ]
: []
}
}
return this.flowProducer.add(lastJob)
}
2022-08-09 09:09:31 +02:00
async createJobWithChildren (parent: CreateJobArgument & CreateJobOptions, children: (CreateJobArgument & CreateJobOptions)[]) {
return this.flowProducer.add({
...this.buildJobFlowOption(parent),
children: children.map(c => this.buildJobFlowOption(c))
})
}
private buildJobFlowOption (job: CreateJobArgument & CreateJobOptions) {
return {
name: 'job',
data: job.payload,
queueName: job.type,
opts: this.buildJobOptions(job.type as JobType, pick(job, [ 'priority', 'delay' ]))
}
}
2022-08-08 15:48:17 +02:00
private buildJobOptions (type: JobType, options: CreateJobOptions = {}): JobsOptions {
return {
2018-07-10 17:02:20 +02:00
backoff: { delay: 60 * 1000, type: 'exponential' },
2022-08-08 15:48:17 +02:00
attempts: JOB_ATTEMPTS[type],
priority: options.priority,
delay: options.delay
2018-07-10 17:02:20 +02:00
}
}
2022-08-08 15:48:17 +02:00
// ---------------------------------------------------------------------------
2019-12-04 14:49:59 +01:00
async listForApi (options: {
2020-12-14 12:00:35 +01:00
state?: JobState
2020-01-31 16:56:52 +01:00
start: number
count: number
asc?: boolean
2019-12-04 14:49:59 +01:00
jobType: JobType
2021-08-27 14:32:44 +02:00
}): Promise<Job[]> {
2020-12-14 12:00:35 +01:00
const { state, start, count, asc, jobType } = options
2022-08-09 11:35:07 +02:00
const states = this.buildStateFilter(state)
const filteredJobTypes = this.buildTypeFilter(jobType)
2022-08-09 11:35:07 +02:00
let results: Job[] = []
2019-12-04 14:49:59 +01:00
for (const jobType of filteredJobTypes) {
2022-08-08 10:42:08 +02:00
const queue: Queue = this.queues[jobType]
2018-07-10 17:02:20 +02:00
if (queue === undefined) {
logger.error('Unknown queue %s to list jobs.', jobType)
continue
}
2020-12-14 12:00:35 +01:00
const jobs = await queue.getJobs(states, 0, start + count, asc)
2018-07-10 17:02:20 +02:00
results = results.concat(jobs)
}
2018-07-10 17:02:20 +02:00
results.sort((j1: any, j2: any) => {
if (j1.timestamp < j2.timestamp) return -1
else if (j1.timestamp === j2.timestamp) return 0
2018-07-10 17:02:20 +02:00
return 1
})
2018-07-10 17:02:20 +02:00
if (asc === false) results.reverse()
2018-07-10 17:02:20 +02:00
return results.slice(start, start + count)
}
2020-12-14 12:00:35 +01:00
async count (state: JobState, jobType?: JobType): Promise<number> {
const states = state ? [ state ] : jobStates
2022-08-09 11:35:07 +02:00
const filteredJobTypes = this.buildTypeFilter(jobType)
2018-02-12 11:25:09 +01:00
2022-08-09 11:35:07 +02:00
let total = 0
2019-12-04 14:49:59 +01:00
for (const type of filteredJobTypes) {
2020-01-31 16:56:52 +01:00
const queue = this.queues[type]
2018-07-10 17:02:20 +02:00
if (queue === undefined) {
logger.error('Unknown queue %s to count jobs.', type)
continue
}
2018-02-12 11:25:09 +01:00
2018-07-10 17:02:20 +02:00
const counts = await queue.getJobCounts()
2018-02-12 11:25:09 +01:00
2020-12-13 19:27:25 +01:00
for (const s of states) {
total += counts[s]
}
2018-07-10 17:02:20 +02:00
}
2018-02-12 11:25:09 +01:00
2018-07-10 17:02:20 +02:00
return total
2018-02-12 11:25:09 +01:00
}
2022-08-09 11:35:07 +02:00
private buildStateFilter (state?: JobState) {
if (!state) return jobStates
const states = [ state ]
// Include parent if filtering on waiting
if (state === 'waiting') states.push('waiting-children')
return states
}
private buildTypeFilter (jobType?: JobType) {
if (!jobType) return jobTypes
return jobTypes.filter(t => t === jobType)
}
2022-07-05 15:43:21 +02:00
async getStats () {
const promises = jobTypes.map(async t => ({ jobType: t, counts: await this.queues[t].getJobCounts() }))
return Promise.all(promises)
}
2022-08-08 15:48:17 +02:00
// ---------------------------------------------------------------------------
async removeOldJobs () {
2018-07-10 17:02:20 +02:00
for (const key of Object.keys(this.queues)) {
2022-08-08 10:42:08 +02:00
const queue: Queue = this.queues[key]
await queue.clean(JOB_COMPLETED_LIFETIME, 100, 'completed')
2018-07-10 17:02:20 +02:00
}
}
2018-08-29 16:26:25 +02:00
private addRepeatableJobs () {
2022-08-08 10:42:08 +02:00
this.queues['videos-views-stats'].add('job', {}, {
repeat: REPEAT_JOBS['videos-views-stats']
2020-01-31 16:56:52 +01:00
}).catch(err => logger.error('Cannot add repeatable job.', { err }))
if (CONFIG.FEDERATION.VIDEOS.CLEANUP_REMOTE_INTERACTIONS) {
2022-08-08 10:42:08 +02:00
this.queues['activitypub-cleaner'].add('job', {}, {
repeat: REPEAT_JOBS['activitypub-cleaner']
}).catch(err => logger.error('Cannot add repeatable job.', { err }))
}
2018-08-29 16:26:25 +02:00
}
private getJobConcurrency (jobType: JobType) {
if (jobType === 'video-transcoding') return CONFIG.TRANSCODING.CONCURRENCY
if (jobType === 'video-import') return CONFIG.IMPORT.VIDEOS.CONCURRENCY
return JOB_CONCURRENCY[jobType]
}
static get Instance () {
return this.instance || (this.instance = new this())
}
}
// ---------------------------------------------------------------------------
export {
2019-12-04 14:49:59 +01:00
jobTypes,
JobQueue
}