PeerTube/server/tests/api/live/live-save-replay.ts

458 lines
15 KiB
TypeScript
Raw Normal View History

2020-11-04 14:16:57 +01:00
/* eslint-disable @typescript-eslint/no-unused-expressions,@typescript-eslint/require-await */
import 'mocha'
import * as chai from 'chai'
import { FfmpegCommand } from 'fluent-ffmpeg'
import { checkLiveCleanup } from '@server/tests/shared'
import { wait } from '@shared/core-utils'
2022-05-03 11:38:07 +02:00
import { HttpStatusCode, LiveVideoCreate, LiveVideoError, VideoPrivacy, VideoState } from '@shared/models'
2020-11-04 14:16:57 +01:00
import {
cleanupTests,
2021-07-07 11:51:09 +02:00
ConfigCommand,
2021-07-16 09:47:51 +02:00
createMultipleServers,
2021-07-16 14:27:30 +02:00
doubleFollow,
findExternalSavedVideo,
2021-07-16 09:47:51 +02:00
PeerTubeServer,
2020-11-04 14:16:57 +01:00
setAccessTokensToServers,
setDefaultVideoChannel,
stopFfmpeg,
testFfmpegStreamError,
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
waitJobs,
waitUntilLivePublishedOnAllServers,
waitUntilLiveReplacedByReplayOnAllServers,
waitUntilLiveWaitingOnAllServers
} from '@shared/server-commands'
2020-11-04 14:16:57 +01:00
const expect = chai.expect
describe('Save replay setting', function () {
2021-07-16 09:47:51 +02:00
let servers: PeerTubeServer[] = []
2020-11-04 14:16:57 +01:00
let liveVideoUUID: string
let ffmpegCommand: FfmpegCommand
async function createLiveWrapper (options: { permanent: boolean, replay: boolean }) {
2020-11-04 14:16:57 +01:00
if (liveVideoUUID) {
try {
2021-07-16 09:04:35 +02:00
await servers[0].videos.remove({ id: liveVideoUUID })
2020-11-04 14:16:57 +01:00
await waitJobs(servers)
} catch {}
}
const attributes: LiveVideoCreate = {
2021-07-16 09:04:35 +02:00
channelId: servers[0].store.channel.id,
2020-11-04 14:16:57 +01:00
privacy: VideoPrivacy.PUBLIC,
name: 'my super live',
saveReplay: options.replay,
permanentLive: options.permanent
2020-11-04 14:16:57 +01:00
}
2021-07-16 09:04:35 +02:00
const { uuid } = await servers[0].live.create({ fields: attributes })
2021-07-08 10:18:40 +02:00
return uuid
2020-11-04 14:16:57 +01:00
}
2022-04-22 09:50:20 +02:00
async function publishLive (options: { permanent: boolean, replay: boolean }) {
liveVideoUUID = await createLiveWrapper(options)
const ffmpegCommand = await servers[0].live.sendRTMPStreamInVideo({ videoId: liveVideoUUID })
await waitUntilLivePublishedOnAllServers(servers, liveVideoUUID)
const liveDetails = await servers[0].videos.get({ id: liveVideoUUID })
await waitJobs(servers)
await checkVideosExist(liveVideoUUID, true, HttpStatusCode.OK_200)
return { ffmpegCommand, liveDetails }
}
async function publishLiveAndDelete (options: { permanent: boolean, replay: boolean }) {
const { ffmpegCommand, liveDetails } = await publishLive(options)
await Promise.all([
servers[0].videos.remove({ id: liveVideoUUID }),
testFfmpegStreamError(ffmpegCommand, true)
])
await waitJobs(servers)
await wait(5000)
await waitJobs(servers)
return { liveDetails }
}
async function publishLiveAndBlacklist (options: { permanent: boolean, replay: boolean }) {
const { ffmpegCommand, liveDetails } = await publishLive(options)
await Promise.all([
servers[0].blacklist.add({ videoId: liveVideoUUID, reason: 'bad live', unfederate: true }),
testFfmpegStreamError(ffmpegCommand, true)
])
await waitJobs(servers)
await wait(5000)
await waitJobs(servers)
return { liveDetails }
}
2021-07-15 10:02:54 +02:00
async function checkVideosExist (videoId: string, existsInList: boolean, expectedStatus?: number) {
2020-11-04 14:16:57 +01:00
for (const server of servers) {
const length = existsInList ? 1 : 0
2021-07-16 09:04:35 +02:00
const { data, total } = await server.videos.list()
2021-07-15 10:02:54 +02:00
expect(data).to.have.lengthOf(length)
expect(total).to.equal(length)
2020-11-04 14:16:57 +01:00
2021-07-15 10:02:54 +02:00
if (expectedStatus) {
2021-07-16 09:04:35 +02:00
await server.videos.get({ id: videoId, expectedStatus })
2020-11-04 14:16:57 +01:00
}
}
}
async function checkVideoState (videoId: string, state: VideoState) {
for (const server of servers) {
2021-07-16 09:04:35 +02:00
const video = await server.videos.get({ id: videoId })
2021-07-15 10:02:54 +02:00
expect(video.state.id).to.equal(state)
2020-11-04 14:16:57 +01:00
}
}
before(async function () {
this.timeout(120000)
2021-07-16 09:47:51 +02:00
servers = await createMultipleServers(2)
2020-11-04 14:16:57 +01:00
// Get the access tokens
await setAccessTokensToServers(servers)
await setDefaultVideoChannel(servers)
// Server 1 and server 2 follow each other
await doubleFollow(servers[0], servers[1])
2021-07-16 09:04:35 +02:00
await servers[0].config.updateCustomSubConfig({
2021-07-07 11:51:09 +02:00
newConfig: {
live: {
enabled: true,
allowReplay: true,
maxDuration: -1,
transcoding: {
enabled: false,
resolutions: ConfigCommand.getCustomConfigResolutions(true)
}
2020-11-04 14:16:57 +01:00
}
}
})
})
describe('With save replay disabled', function () {
2022-05-03 11:38:07 +02:00
let sessionStartDateMin: Date
let sessionStartDateMax: Date
let sessionEndDateMin: Date
2020-11-04 14:16:57 +01:00
it('Should correctly create and federate the "waiting for stream" live', async function () {
this.timeout(20000)
liveVideoUUID = await createLiveWrapper({ permanent: false, replay: false })
2020-11-04 14:16:57 +01:00
await waitJobs(servers)
await checkVideosExist(liveVideoUUID, false, HttpStatusCode.OK_200)
2020-11-04 14:16:57 +01:00
await checkVideoState(liveVideoUUID, VideoState.WAITING_FOR_LIVE)
})
it('Should correctly have updated the live and federated it when streaming in the live', async function () {
2021-02-18 14:44:12 +01:00
this.timeout(30000)
2020-11-04 14:16:57 +01:00
2021-07-16 09:04:35 +02:00
ffmpegCommand = await servers[0].live.sendRTMPStreamInVideo({ videoId: liveVideoUUID })
2021-02-18 14:44:12 +01:00
2022-05-03 11:38:07 +02:00
sessionStartDateMin = new Date()
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
await waitUntilLivePublishedOnAllServers(servers, liveVideoUUID)
2022-05-03 11:38:07 +02:00
sessionStartDateMax = new Date()
2020-11-04 14:16:57 +01:00
await waitJobs(servers)
await checkVideosExist(liveVideoUUID, true, HttpStatusCode.OK_200)
2020-11-04 14:16:57 +01:00
await checkVideoState(liveVideoUUID, VideoState.PUBLISHED)
})
it('Should correctly delete the video files after the stream ended', async function () {
2020-12-11 10:36:05 +01:00
this.timeout(40000)
2020-11-04 14:16:57 +01:00
2022-05-03 11:38:07 +02:00
sessionEndDateMin = new Date()
2020-11-04 14:16:57 +01:00
await stopFfmpeg(ffmpegCommand)
2021-02-18 14:44:12 +01:00
for (const server of servers) {
2021-07-16 09:04:35 +02:00
await server.live.waitUntilEnded({ videoId: liveVideoUUID })
2021-02-18 14:44:12 +01:00
}
2020-11-04 14:16:57 +01:00
await waitJobs(servers)
// Live still exist, but cannot be played anymore
await checkVideosExist(liveVideoUUID, false, HttpStatusCode.OK_200)
2020-11-04 14:16:57 +01:00
await checkVideoState(liveVideoUUID, VideoState.LIVE_ENDED)
// No resolutions saved since we did not save replay
await checkLiveCleanup(servers[0], liveVideoUUID, [])
2020-11-04 14:16:57 +01:00
})
2022-05-03 11:38:07 +02:00
it('Should have appropriate ended session', async function () {
const { data, total } = await servers[0].live.listSessions({ videoId: liveVideoUUID })
expect(total).to.equal(1)
expect(data).to.have.lengthOf(1)
const session = data[0]
const startDate = new Date(session.startDate)
expect(startDate).to.be.above(sessionStartDateMin)
expect(startDate).to.be.below(sessionStartDateMax)
expect(session.endDate).to.exist
expect(new Date(session.endDate)).to.be.above(sessionEndDateMin)
expect(session.saveReplay).to.be.false
2022-05-03 11:38:07 +02:00
expect(session.error).to.not.exist
expect(session.replayVideo).to.not.exist
})
2020-11-04 14:16:57 +01:00
it('Should correctly terminate the stream on blacklist and delete the live', async function () {
this.timeout(40000)
2022-04-22 09:50:20 +02:00
await publishLiveAndBlacklist({ permanent: false, replay: false })
2020-11-04 14:16:57 +01:00
await checkVideosExist(liveVideoUUID, false)
2021-07-16 09:04:35 +02:00
await servers[0].videos.get({ id: liveVideoUUID, expectedStatus: HttpStatusCode.UNAUTHORIZED_401 })
await servers[1].videos.get({ id: liveVideoUUID, expectedStatus: HttpStatusCode.NOT_FOUND_404 })
2020-11-04 14:16:57 +01:00
2021-02-19 14:30:00 +01:00
await wait(5000)
await waitJobs(servers)
await checkLiveCleanup(servers[0], liveVideoUUID, [])
2020-11-04 14:16:57 +01:00
})
2022-05-03 11:38:07 +02:00
it('Should have blacklisted session error', async function () {
const session = await servers[0].live.findLatestSession({ videoId: liveVideoUUID })
expect(session.startDate).to.exist
expect(session.endDate).to.exist
expect(session.error).to.equal(LiveVideoError.BLACKLISTED)
expect(session.replayVideo).to.not.exist
})
2020-11-04 14:16:57 +01:00
it('Should correctly terminate the stream on delete and delete the video', async function () {
this.timeout(40000)
2022-04-22 09:50:20 +02:00
await publishLiveAndDelete({ permanent: false, replay: false })
2020-11-04 14:16:57 +01:00
await checkVideosExist(liveVideoUUID, false, HttpStatusCode.NOT_FOUND_404)
await checkLiveCleanup(servers[0], liveVideoUUID, [])
2020-11-04 14:16:57 +01:00
})
})
describe('With save replay enabled on non permanent live', function () {
2020-11-04 14:16:57 +01:00
it('Should correctly create and federate the "waiting for stream" live', async function () {
this.timeout(20000)
liveVideoUUID = await createLiveWrapper({ permanent: false, replay: true })
2020-11-04 14:16:57 +01:00
await waitJobs(servers)
await checkVideosExist(liveVideoUUID, false, HttpStatusCode.OK_200)
2020-11-04 14:16:57 +01:00
await checkVideoState(liveVideoUUID, VideoState.WAITING_FOR_LIVE)
})
it('Should correctly have updated the live and federated it when streaming in the live', async function () {
this.timeout(20000)
2021-07-16 09:04:35 +02:00
ffmpegCommand = await servers[0].live.sendRTMPStreamInVideo({ videoId: liveVideoUUID })
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
await waitUntilLivePublishedOnAllServers(servers, liveVideoUUID)
2020-11-04 14:16:57 +01:00
await waitJobs(servers)
await checkVideosExist(liveVideoUUID, true, HttpStatusCode.OK_200)
2020-11-04 14:16:57 +01:00
await checkVideoState(liveVideoUUID, VideoState.PUBLISHED)
})
it('Should correctly have saved the live and federated it after the streaming', async function () {
this.timeout(30000)
const session = await servers[0].live.findLatestSession({ videoId: liveVideoUUID })
expect(session.endDate).to.not.exist
expect(session.endingProcessed).to.be.false
expect(session.saveReplay).to.be.true
2020-11-04 14:16:57 +01:00
await stopFfmpeg(ffmpegCommand)
await waitUntilLiveReplacedByReplayOnAllServers(servers, liveVideoUUID)
2020-11-04 14:16:57 +01:00
await waitJobs(servers)
// Live has been transcoded
await checkVideosExist(liveVideoUUID, true, HttpStatusCode.OK_200)
2020-11-04 14:16:57 +01:00
await checkVideoState(liveVideoUUID, VideoState.PUBLISHED)
})
2022-05-03 11:38:07 +02:00
it('Should find the replay live session', async function () {
const session = await servers[0].live.getReplaySession({ videoId: liveVideoUUID })
expect(session).to.exist
expect(session.startDate).to.exist
expect(session.endDate).to.exist
expect(session.error).to.not.exist
expect(session.saveReplay).to.be.true
expect(session.endingProcessed).to.be.true
2022-05-03 11:38:07 +02:00
expect(session.replayVideo).to.exist
expect(session.replayVideo.id).to.exist
expect(session.replayVideo.shortUUID).to.exist
expect(session.replayVideo.uuid).to.equal(liveVideoUUID)
})
2020-11-04 14:16:57 +01:00
it('Should update the saved live and correctly federate the updated attributes', async function () {
this.timeout(30000)
2021-07-16 09:04:35 +02:00
await servers[0].videos.update({ id: liveVideoUUID, attributes: { name: 'video updated' } })
2020-11-04 14:16:57 +01:00
await waitJobs(servers)
for (const server of servers) {
2021-07-16 09:04:35 +02:00
const video = await server.videos.get({ id: liveVideoUUID })
2021-07-15 10:02:54 +02:00
expect(video.name).to.equal('video updated')
expect(video.isLive).to.be.false
2020-11-04 14:16:57 +01:00
}
})
it('Should have cleaned up the live files', async function () {
await checkLiveCleanup(servers[0], liveVideoUUID, [ 720 ])
2020-11-04 14:16:57 +01:00
})
it('Should correctly terminate the stream on blacklist and blacklist the saved replay video', async function () {
2022-07-08 11:51:41 +02:00
this.timeout(120000)
2020-11-04 14:16:57 +01:00
2022-04-22 09:50:20 +02:00
await publishLiveAndBlacklist({ permanent: false, replay: true })
2020-11-04 14:16:57 +01:00
await checkVideosExist(liveVideoUUID, false)
2021-07-16 09:04:35 +02:00
await servers[0].videos.get({ id: liveVideoUUID, expectedStatus: HttpStatusCode.UNAUTHORIZED_401 })
await servers[1].videos.get({ id: liveVideoUUID, expectedStatus: HttpStatusCode.NOT_FOUND_404 })
2020-11-04 14:16:57 +01:00
2021-02-19 14:30:00 +01:00
await wait(5000)
await waitJobs(servers)
await checkLiveCleanup(servers[0], liveVideoUUID, [ 720 ])
2020-11-04 14:16:57 +01:00
})
it('Should correctly terminate the stream on delete and delete the video', async function () {
this.timeout(40000)
2022-04-22 09:50:20 +02:00
await publishLiveAndDelete({ permanent: false, replay: true })
2020-11-04 14:16:57 +01:00
await checkVideosExist(liveVideoUUID, false, HttpStatusCode.NOT_FOUND_404)
await checkLiveCleanup(servers[0], liveVideoUUID, [])
})
})
describe('With save replay enabled on permanent live', function () {
let lastReplayUUID: string
it('Should correctly create and federate the "waiting for stream" live', async function () {
this.timeout(20000)
liveVideoUUID = await createLiveWrapper({ permanent: true, replay: true })
await waitJobs(servers)
await checkVideosExist(liveVideoUUID, false, HttpStatusCode.OK_200)
await checkVideoState(liveVideoUUID, VideoState.WAITING_FOR_LIVE)
})
it('Should correctly have updated the live and federated it when streaming in the live', async function () {
this.timeout(20000)
ffmpegCommand = await servers[0].live.sendRTMPStreamInVideo({ videoId: liveVideoUUID })
await waitUntilLivePublishedOnAllServers(servers, liveVideoUUID)
await waitJobs(servers)
await checkVideosExist(liveVideoUUID, true, HttpStatusCode.OK_200)
await checkVideoState(liveVideoUUID, VideoState.PUBLISHED)
})
it('Should correctly have saved the live and federated it after the streaming', async function () {
this.timeout(30000)
const liveDetails = await servers[0].videos.get({ id: liveVideoUUID })
await stopFfmpeg(ffmpegCommand)
await waitUntilLiveWaitingOnAllServers(servers, liveVideoUUID)
await waitJobs(servers)
const video = await findExternalSavedVideo(servers[0], liveDetails)
expect(video).to.exist
for (const server of servers) {
await server.videos.get({ id: video.uuid })
}
lastReplayUUID = video.uuid
})
2022-05-03 11:38:07 +02:00
it('Should have appropriate ended session and replay live session', async function () {
const { data, total } = await servers[0].live.listSessions({ videoId: liveVideoUUID })
expect(total).to.equal(1)
expect(data).to.have.lengthOf(1)
const sessionFromLive = data[0]
const sessionFromReplay = await servers[0].live.getReplaySession({ videoId: lastReplayUUID })
for (const session of [ sessionFromLive, sessionFromReplay ]) {
expect(session.startDate).to.exist
expect(session.endDate).to.exist
expect(session.error).to.not.exist
expect(session.replayVideo).to.exist
expect(session.replayVideo.id).to.exist
expect(session.replayVideo.shortUUID).to.exist
expect(session.replayVideo.uuid).to.equal(lastReplayUUID)
}
})
it('Should have cleaned up the live files', async function () {
await checkLiveCleanup(servers[0], liveVideoUUID, [])
})
it('Should correctly terminate the stream on blacklist and blacklist the saved replay video', async function () {
2022-07-08 11:51:41 +02:00
this.timeout(120000)
await servers[0].videos.remove({ id: lastReplayUUID })
2022-04-22 09:50:20 +02:00
const { liveDetails } = await publishLiveAndBlacklist({ permanent: true, replay: true })
const replay = await findExternalSavedVideo(servers[0], liveDetails)
expect(replay).to.exist
for (const videoId of [ liveVideoUUID, replay.uuid ]) {
await checkVideosExist(videoId, false)
await servers[0].videos.get({ id: videoId, expectedStatus: HttpStatusCode.UNAUTHORIZED_401 })
await servers[1].videos.get({ id: videoId, expectedStatus: HttpStatusCode.NOT_FOUND_404 })
}
await checkLiveCleanup(servers[0], liveVideoUUID, [])
})
it('Should correctly terminate the stream on delete and not save the video', async function () {
this.timeout(40000)
2022-04-22 09:50:20 +02:00
const { liveDetails } = await publishLiveAndDelete({ permanent: true, replay: true })
const replay = await findExternalSavedVideo(servers[0], liveDetails)
expect(replay).to.not.exist
await checkVideosExist(liveVideoUUID, false, HttpStatusCode.NOT_FOUND_404)
await checkLiveCleanup(servers[0], liveVideoUUID, [])
2020-11-04 14:16:57 +01:00
})
})
after(async function () {
await cleanupTests(servers)
})
})