PeerTube/server/tests/api/redundancy/redundancy.ts

749 lines
21 KiB
TypeScript
Raw Normal View History

2020-01-31 16:56:52 +01:00
/* eslint-disable @typescript-eslint/no-unused-expressions,@typescript-eslint/require-await */
import 'mocha'
2021-02-01 11:57:21 +01:00
import * as chai from 'chai'
import { readdir } from 'fs-extra'
2021-08-27 14:32:44 +02:00
import magnetUtil from 'magnet-uri'
2021-07-22 14:28:03 +02:00
import { basename, join } from 'path'
import {
2019-02-28 11:14:26 +01:00
checkSegmentHash,
2020-01-10 10:11:28 +01:00
checkVideoFilesWereRemoved,
cleanupTests,
2021-07-16 09:47:51 +02:00
createMultipleServers,
2021-07-16 14:27:30 +02:00
doubleFollow,
2019-02-28 11:14:26 +01:00
killallServers,
2021-07-22 14:28:03 +02:00
makeRawRequest,
2021-07-16 09:47:51 +02:00
PeerTubeServer,
2021-07-16 14:27:30 +02:00
root,
2021-07-22 14:28:03 +02:00
saveVideoInServers,
2019-02-28 11:14:26 +01:00
setAccessTokensToServers,
wait,
2021-07-13 09:43:59 +02:00
waitJobs
2021-07-07 10:56:45 +02:00
} from '@shared/extra-utils'
2021-07-22 14:28:03 +02:00
import {
HttpStatusCode,
VideoDetails,
VideoFile,
VideoPrivacy,
VideoRedundancyStrategy,
VideoRedundancyStrategyWithManual
} from '@shared/models'
const expect = chai.expect
2021-07-16 09:47:51 +02:00
let servers: PeerTubeServer[] = []
2021-07-22 14:28:03 +02:00
let video1Server2: VideoDetails
2021-07-22 14:28:03 +02:00
async function checkMagnetWebseeds (file: VideoFile, baseWebseeds: string[], server: PeerTubeServer) {
const parsed = magnetUtil.decode(file.magnetUri)
for (const ws of baseWebseeds) {
2021-07-22 14:28:03 +02:00
const found = parsed.urlList.find(url => url === `${ws}${basename(file.fileUrl)}`)
expect(found, `Webseed ${ws} not found in ${file.magnetUri} on server ${server.url}`).to.not.be.undefined
}
expect(parsed.urlList).to.have.lengthOf(baseWebseeds.length)
2021-07-22 14:28:03 +02:00
for (const url of parsed.urlList) {
await makeRawRequest(url, HttpStatusCode.OK_200)
}
}
async function createServers (strategy: VideoRedundancyStrategy | null, additionalParams: any = {}, withWebtorrent = true) {
2020-01-10 10:11:28 +01:00
const strategies: any[] = []
if (strategy !== null) {
strategies.push(
2021-07-13 09:43:59 +02:00
{
2020-01-10 10:11:28 +01:00
min_lifetime: '1 hour',
strategy: strategy,
2021-07-13 09:43:59 +02:00
size: '400KB',
...additionalParams
}
2020-01-10 10:11:28 +01:00
)
}
const config = {
2019-01-29 08:37:25 +01:00
transcoding: {
2021-02-01 11:18:50 +01:00
webtorrent: {
enabled: withWebtorrent
},
2019-01-29 08:37:25 +01:00
hls: {
enabled: true
}
},
redundancy: {
videos: {
check_interval: '5 seconds',
2020-01-10 10:11:28 +01:00
strategies
}
}
}
2020-01-10 10:11:28 +01:00
2021-07-16 09:47:51 +02:00
servers = await createMultipleServers(3, config)
// Get the access tokens
await setAccessTokensToServers(servers)
{
2021-07-22 14:28:03 +02:00
const { id } = await servers[1].videos.upload({ attributes: { name: 'video 1 server 2' } })
video1Server2 = await servers[1].videos.get({ id })
2021-07-22 14:28:03 +02:00
await servers[1].videos.view({ id })
}
await waitJobs(servers)
// Server 1 and server 2 follow each other
2020-01-31 16:56:52 +01:00
await doubleFollow(servers[0], servers[1])
// Server 1 and server 3 follow each other
2020-01-31 16:56:52 +01:00
await doubleFollow(servers[0], servers[2])
// Server 2 and server 3 follow each other
2020-01-31 16:56:52 +01:00
await doubleFollow(servers[1], servers[2])
await waitJobs(servers)
}
2021-07-22 14:28:03 +02:00
async function ensureSameFilenames (videoUUID: string) {
let webtorrentFilenames: string[]
let hlsFilenames: string[]
for (const server of servers) {
const video = await server.videos.getWithToken({ id: videoUUID })
// Ensure we use the same filenames that the origin
const localWebtorrentFilenames = video.files.map(f => basename(f.fileUrl)).sort()
const localHLSFilenames = video.streamingPlaylists[0].files.map(f => basename(f.fileUrl)).sort()
if (webtorrentFilenames) expect(webtorrentFilenames).to.deep.equal(localWebtorrentFilenames)
else webtorrentFilenames = localWebtorrentFilenames
if (hlsFilenames) expect(hlsFilenames).to.deep.equal(localHLSFilenames)
else hlsFilenames = localHLSFilenames
}
return { webtorrentFilenames, hlsFilenames }
}
2019-01-29 08:37:25 +01:00
async function check1WebSeed (videoUUID?: string) {
2021-07-22 14:28:03 +02:00
if (!videoUUID) videoUUID = video1Server2.uuid
const webseeds = [
2021-07-22 14:28:03 +02:00
`http://localhost:${servers[1].port}/static/webseed/`
]
for (const server of servers) {
2019-01-29 08:37:25 +01:00
// With token to avoid issues with video follow constraints
2021-07-16 09:04:35 +02:00
const video = await server.videos.getWithToken({ id: videoUUID })
2019-01-29 08:37:25 +01:00
for (const f of video.files) {
2021-07-22 14:28:03 +02:00
await checkMagnetWebseeds(f, webseeds, server)
}
}
2021-07-22 14:28:03 +02:00
await ensureSameFilenames(videoUUID)
}
2019-01-29 08:37:25 +01:00
async function check2Webseeds (videoUUID?: string) {
2021-07-22 14:28:03 +02:00
if (!videoUUID) videoUUID = video1Server2.uuid
const webseeds = [
2021-07-22 14:28:03 +02:00
`http://localhost:${servers[0].port}/static/redundancy/`,
`http://localhost:${servers[1].port}/static/webseed/`
]
for (const server of servers) {
2021-07-16 09:04:35 +02:00
const video = await server.videos.get({ id: videoUUID })
for (const file of video.files) {
2021-07-22 14:28:03 +02:00
await checkMagnetWebseeds(file, webseeds, server)
}
}
2021-07-22 14:28:03 +02:00
const { webtorrentFilenames } = await ensureSameFilenames(videoUUID)
const directories = [
'test' + servers[0].internalServerNumber + '/redundancy',
'test' + servers[1].internalServerNumber + '/videos'
]
for (const directory of directories) {
2018-12-04 17:08:55 +01:00
const files = await readdir(join(root(), directory))
expect(files).to.have.length.at.least(4)
2021-07-22 14:28:03 +02:00
// Ensure we files exist on disk
expect(files.find(f => webtorrentFilenames.includes(f))).to.exist
}
}
2019-01-29 08:37:25 +01:00
async function check0PlaylistRedundancies (videoUUID?: string) {
2021-07-22 14:28:03 +02:00
if (!videoUUID) videoUUID = video1Server2.uuid
2019-01-29 08:37:25 +01:00
for (const server of servers) {
// With token to avoid issues with video follow constraints
2021-07-16 09:04:35 +02:00
const video = await server.videos.getWithToken({ id: videoUUID })
2019-01-29 08:37:25 +01:00
expect(video.streamingPlaylists).to.be.an('array')
expect(video.streamingPlaylists).to.have.lengthOf(1)
expect(video.streamingPlaylists[0].redundancies).to.have.lengthOf(0)
}
2021-07-22 14:28:03 +02:00
await ensureSameFilenames(videoUUID)
2019-01-29 08:37:25 +01:00
}
async function check1PlaylistRedundancies (videoUUID?: string) {
2021-07-22 14:28:03 +02:00
if (!videoUUID) videoUUID = video1Server2.uuid
2019-01-29 08:37:25 +01:00
for (const server of servers) {
2021-07-16 09:04:35 +02:00
const video = await server.videos.get({ id: videoUUID })
2019-01-29 08:37:25 +01:00
expect(video.streamingPlaylists).to.have.lengthOf(1)
expect(video.streamingPlaylists[0].redundancies).to.have.lengthOf(1)
const redundancy = video.streamingPlaylists[0].redundancies[0]
expect(redundancy.baseUrl).to.equal(servers[0].url + '/static/redundancy/hls/' + videoUUID)
}
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
const baseUrlPlaylist = servers[1].url + '/static/streaming-playlists/hls/' + videoUUID
const baseUrlSegment = servers[0].url + '/static/redundancy/hls/' + videoUUID
2021-07-16 09:04:35 +02:00
const video = await servers[0].videos.get({ id: videoUUID })
2021-07-15 10:02:54 +02:00
const hlsPlaylist = video.streamingPlaylists[0]
for (const resolution of [ 240, 360, 480, 720 ]) {
Add support for saving video files to object storage (#4290) * Add support for saving video files to object storage * Add support for custom url generation on s3 stored files Uses two config keys to support url generation that doesn't directly go to (compatible s3). Can be used to generate urls to any cache server or CDN. * Upload files to s3 concurrently and delete originals afterwards * Only publish after move to object storage is complete * Use base url instead of url template * Fix mistyped config field * Add rudenmentary way to download before transcode * Implement Chocobozzz suggestions https://github.com/Chocobozzz/PeerTube/pull/4290#issuecomment-891670478 The remarks in question: Try to use objectStorage prefix instead of s3 prefix for your function/variables/config names Prefer to use a tree for the config: s3.streaming_playlists_bucket -> object_storage.streaming_playlists.bucket Use uppercase for config: S3.STREAMING_PLAYLISTS_BUCKETINFO.bucket -> OBJECT_STORAGE.STREAMING_PLAYLISTS.BUCKET (maybe BUCKET_NAME instead of BUCKET) I suggest to rename moveJobsRunning to pendingMovingJobs (or better, create a dedicated videoJobInfo table with a pendingMove & videoId columns so we could also use this table to track pending transcoding jobs) https://github.com/Chocobozzz/PeerTube/pull/4290/files#diff-3e26d41ca4bda1de8e1747af70ca2af642abcc1e9e0bfb94239ff2165acfbde5R19 uses a string instead of an integer I think we should store the origin object storage URL in fileUrl, without base_url injection. Instead, inject the base_url at "runtime" so admins can easily change this configuration without running a script to update DB URLs * Import correct function * Support multipart upload * Remove import of node 15.0 module stream/promises * Extend maximum upload job length Using the same value as for redundancy downloading seems logical * Use dynamic part size for really large uploads Also adds very small part size for local testing * Fix decreasePendingMove query * Resolve various PR comments * Move to object storage after optimize * Make upload size configurable and increase default * Prune webtorrent files that are stored in object storage * Move files after transcoding jobs * Fix federation * Add video path manager * Support move to external storage job in client * Fix live object storage tests Co-authored-by: Chocobozzz <me@florianbigard.com>
2021-08-17 08:26:20 +02:00
await checkSegmentHash({ server: servers[1], baseUrlPlaylist, baseUrlSegment, resolution, hlsPlaylist })
}
2019-01-29 08:37:25 +01:00
2021-07-22 14:28:03 +02:00
const { hlsFilenames } = await ensureSameFilenames(videoUUID)
const directories = [
'test' + servers[0].internalServerNumber + '/redundancy/hls',
'test' + servers[1].internalServerNumber + '/streaming-playlists/hls'
]
for (const directory of directories) {
2019-01-29 08:37:25 +01:00
const files = await readdir(join(root(), directory, videoUUID))
expect(files).to.have.length.at.least(4)
2021-07-22 14:28:03 +02:00
// Ensure we files exist on disk
expect(files.find(f => hlsFilenames.includes(f))).to.exist
2019-01-29 08:37:25 +01:00
}
}
2020-01-10 10:11:28 +01:00
async function checkStatsGlobal (strategy: VideoRedundancyStrategyWithManual) {
let totalSize: number = null
let statsLength = 1
if (strategy !== 'manual') {
totalSize = 409600
statsLength = 2
}
2021-07-16 09:04:35 +02:00
const data = await servers[0].stats.get()
2020-01-10 10:11:28 +01:00
expect(data.videosRedundancy).to.have.lengthOf(statsLength)
2019-01-29 08:37:25 +01:00
2020-01-10 10:11:28 +01:00
const stat = data.videosRedundancy[0]
2019-01-29 08:37:25 +01:00
expect(stat.strategy).to.equal(strategy)
2020-01-10 10:11:28 +01:00
expect(stat.totalSize).to.equal(totalSize)
return stat
}
2021-02-01 11:57:21 +01:00
async function checkStatsWith1Redundancy (strategy: VideoRedundancyStrategyWithManual, onlyHls = false) {
2020-01-10 10:11:28 +01:00
const stat = await checkStatsGlobal(strategy)
2019-08-13 10:22:54 +02:00
expect(stat.totalUsed).to.be.at.least(1).and.below(409601)
2021-02-01 11:57:21 +01:00
expect(stat.totalVideoFiles).to.equal(onlyHls ? 4 : 8)
2019-01-29 08:37:25 +01:00
expect(stat.totalVideos).to.equal(1)
}
2021-02-01 11:18:50 +01:00
async function checkStatsWithoutRedundancy (strategy: VideoRedundancyStrategyWithManual) {
2020-01-10 10:11:28 +01:00
const stat = await checkStatsGlobal(strategy)
2019-01-29 08:37:25 +01:00
expect(stat.totalUsed).to.equal(0)
expect(stat.totalVideoFiles).to.equal(0)
expect(stat.totalVideos).to.equal(0)
}
2021-07-07 09:16:40 +02:00
async function findServerFollows () {
2021-07-16 09:04:35 +02:00
const body = await servers[0].follows.getFollowings({ start: 0, count: 5, sort: '-createdAt' })
2021-07-07 09:16:40 +02:00
const follows = body.data
const server2 = follows.find(f => f.following.host === `localhost:${servers[1].port}`)
const server3 = follows.find(f => f.following.host === `localhost:${servers[2].port}`)
return { server2, server3 }
}
async function enableRedundancyOnServer1 () {
2021-07-16 09:04:35 +02:00
await servers[0].redundancy.updateRedundancy({ host: servers[1].host, redundancyAllowed: true })
2021-07-07 09:16:40 +02:00
const { server2, server3 } = await findServerFollows()
expect(server3).to.not.be.undefined
expect(server3.following.hostRedundancyAllowed).to.be.false
expect(server2).to.not.be.undefined
expect(server2.following.hostRedundancyAllowed).to.be.true
}
async function disableRedundancyOnServer1 () {
2021-07-16 09:04:35 +02:00
await servers[0].redundancy.updateRedundancy({ host: servers[1].host, redundancyAllowed: false })
2021-07-07 09:16:40 +02:00
const { server2, server3 } = await findServerFollows()
expect(server3).to.not.be.undefined
expect(server3.following.hostRedundancyAllowed).to.be.false
expect(server2).to.not.be.undefined
expect(server2.following.hostRedundancyAllowed).to.be.false
}
describe('Test videos redundancy', function () {
describe('With most-views strategy', function () {
const strategy = 'most-views'
before(function () {
this.timeout(120000)
return createServers(strategy)
})
it('Should have 1 webseed on the first video', async function () {
2019-01-29 08:37:25 +01:00
await check1WebSeed()
await check0PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWithoutRedundancy(strategy)
})
it('Should enable redundancy on server 1', function () {
return enableRedundancyOnServer1()
})
2018-11-15 11:20:23 +01:00
it('Should have 2 webseeds on the first video', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
await waitJobs(servers)
2021-07-16 09:04:35 +02:00
await servers[0].servers.waitUntilLog('Duplicated ', 5)
await waitJobs(servers)
2019-01-29 08:37:25 +01:00
await check2Webseeds()
await check1PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWith1Redundancy(strategy)
})
it('Should undo redundancy on server 1 and remove duplicated videos', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
await disableRedundancyOnServer1()
await waitJobs(servers)
await wait(5000)
2019-01-29 08:37:25 +01:00
await check1WebSeed()
await check0PlaylistRedundancies()
2021-07-22 14:28:03 +02:00
await checkVideoFilesWereRemoved({ server: servers[0], video: video1Server2, onlyVideoFiles: true })
})
2019-04-24 15:10:37 +02:00
after(async function () {
return cleanupTests(servers)
})
})
describe('With trending strategy', function () {
const strategy = 'trending'
before(function () {
this.timeout(120000)
return createServers(strategy)
})
it('Should have 1 webseed on the first video', async function () {
2019-01-29 08:37:25 +01:00
await check1WebSeed()
await check0PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWithoutRedundancy(strategy)
})
it('Should enable redundancy on server 1', function () {
return enableRedundancyOnServer1()
})
2018-11-15 11:20:23 +01:00
it('Should have 2 webseeds on the first video', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
await waitJobs(servers)
2021-07-16 09:04:35 +02:00
await servers[0].servers.waitUntilLog('Duplicated ', 5)
await waitJobs(servers)
2019-01-29 08:37:25 +01:00
await check2Webseeds()
await check1PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWith1Redundancy(strategy)
})
it('Should unfollow server 3 and keep duplicated videos', async function () {
this.timeout(80000)
await servers[0].follows.unfollow({ target: servers[2] })
await waitJobs(servers)
await wait(5000)
await check2Webseeds()
await check1PlaylistRedundancies()
await checkStatsWith1Redundancy(strategy)
})
it('Should unfollow server 2 and remove duplicated videos', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
2021-07-16 09:04:35 +02:00
await servers[0].follows.unfollow({ target: servers[1] })
await waitJobs(servers)
await wait(5000)
2019-01-29 08:37:25 +01:00
await check1WebSeed()
await check0PlaylistRedundancies()
2021-07-22 14:28:03 +02:00
await checkVideoFilesWereRemoved({ server: servers[0], video: video1Server2, onlyVideoFiles: true })
})
2019-04-24 15:10:37 +02:00
after(async function () {
await cleanupTests(servers)
})
})
describe('With recently added strategy', function () {
const strategy = 'recently-added'
before(function () {
this.timeout(120000)
return createServers(strategy, { min_views: 3 })
})
it('Should have 1 webseed on the first video', async function () {
2019-01-29 08:37:25 +01:00
await check1WebSeed()
await check0PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWithoutRedundancy(strategy)
})
it('Should enable redundancy on server 1', function () {
return enableRedundancyOnServer1()
})
it('Should still have 1 webseed on the first video', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
await waitJobs(servers)
await wait(15000)
await waitJobs(servers)
2019-01-29 08:37:25 +01:00
await check1WebSeed()
await check0PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWithoutRedundancy(strategy)
})
it('Should view 2 times the first video to have > min_views config', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
2021-07-22 14:28:03 +02:00
await servers[0].videos.view({ id: video1Server2.uuid })
await servers[2].videos.view({ id: video1Server2.uuid })
await wait(10000)
await waitJobs(servers)
})
2018-11-15 11:20:23 +01:00
it('Should have 2 webseeds on the first video', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
await waitJobs(servers)
2021-07-16 09:04:35 +02:00
await servers[0].servers.waitUntilLog('Duplicated ', 5)
await waitJobs(servers)
2019-01-29 08:37:25 +01:00
await check2Webseeds()
await check1PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWith1Redundancy(strategy)
})
it('Should remove the video and the redundancy files', async function () {
this.timeout(20000)
2021-07-22 14:28:03 +02:00
await saveVideoInServers(servers, video1Server2.uuid)
await servers[1].videos.remove({ id: video1Server2.uuid })
await waitJobs(servers)
for (const server of servers) {
2021-07-22 14:28:03 +02:00
await checkVideoFilesWereRemoved({ server, video: server.store.videoDetails })
}
})
2019-04-24 15:10:37 +02:00
after(async function () {
await cleanupTests(servers)
})
})
2021-02-01 11:18:50 +01:00
describe('With only HLS files', function () {
const strategy = 'recently-added'
before(async function () {
this.timeout(120000)
await createServers(strategy, { min_views: 3 }, false)
2021-02-01 11:18:50 +01:00
})
it('Should have 0 playlist redundancy on the first video', async function () {
await check1WebSeed()
await check0PlaylistRedundancies()
})
it('Should enable redundancy on server 1', function () {
return enableRedundancyOnServer1()
})
it('Should still have 0 redundancy on the first video', async function () {
this.timeout(80000)
await waitJobs(servers)
await wait(15000)
await waitJobs(servers)
await check0PlaylistRedundancies()
await checkStatsWithoutRedundancy(strategy)
})
it('Should have 1 redundancy on the first video', async function () {
this.timeout(160000)
2021-07-22 14:28:03 +02:00
await servers[0].videos.view({ id: video1Server2.uuid })
await servers[2].videos.view({ id: video1Server2.uuid })
2021-02-01 11:18:50 +01:00
await wait(10000)
await waitJobs(servers)
await waitJobs(servers)
2021-07-16 09:04:35 +02:00
await servers[0].servers.waitUntilLog('Duplicated ', 1)
2021-02-01 11:18:50 +01:00
await waitJobs(servers)
await check1PlaylistRedundancies()
2021-02-01 11:57:21 +01:00
await checkStatsWith1Redundancy(strategy, true)
2021-02-01 11:18:50 +01:00
})
it('Should remove the video and the redundancy files', async function () {
this.timeout(20000)
2021-07-22 14:28:03 +02:00
await saveVideoInServers(servers, video1Server2.uuid)
await servers[1].videos.remove({ id: video1Server2.uuid })
2021-02-01 11:18:50 +01:00
await waitJobs(servers)
for (const server of servers) {
2021-07-22 14:28:03 +02:00
await checkVideoFilesWereRemoved({ server, video: server.store.videoDetails })
2021-02-01 11:18:50 +01:00
}
})
2021-02-01 11:57:21 +01:00
after(async function () {
await cleanupTests(servers)
})
2021-02-01 11:18:50 +01:00
})
2020-01-10 10:11:28 +01:00
describe('With manual strategy', function () {
before(function () {
this.timeout(120000)
return createServers(null)
2020-01-10 10:11:28 +01:00
})
it('Should have 1 webseed on the first video', async function () {
await check1WebSeed()
await check0PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWithoutRedundancy('manual')
2020-01-10 10:11:28 +01:00
})
it('Should create a redundancy on first video', async function () {
2021-07-22 14:28:03 +02:00
await servers[0].redundancy.addVideo({ videoId: video1Server2.id })
2020-01-10 10:11:28 +01:00
})
it('Should have 2 webseeds on the first video', async function () {
this.timeout(80000)
await waitJobs(servers)
2021-07-16 09:04:35 +02:00
await servers[0].servers.waitUntilLog('Duplicated ', 5)
2020-01-10 10:11:28 +01:00
await waitJobs(servers)
await check2Webseeds()
await check1PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWith1Redundancy('manual')
2020-01-10 10:11:28 +01:00
})
it('Should manually remove redundancies on server 1 and remove duplicated videos', async function () {
this.timeout(80000)
2021-07-16 09:04:35 +02:00
const body = await servers[0].redundancy.listVideos({ target: 'remote-videos' })
2020-01-10 10:11:28 +01:00
2021-07-07 10:56:45 +02:00
const videos = body.data
2020-01-10 10:11:28 +01:00
expect(videos).to.have.lengthOf(1)
const video = videos[0]
2021-07-07 10:56:45 +02:00
2020-01-10 10:11:28 +01:00
for (const r of video.redundancies.files.concat(video.redundancies.streamingPlaylists)) {
2021-07-16 09:04:35 +02:00
await servers[0].redundancy.removeVideo({ redundancyId: r.id })
2020-01-10 10:11:28 +01:00
}
await waitJobs(servers)
await wait(5000)
await check1WebSeed()
await check0PlaylistRedundancies()
2021-07-22 14:28:03 +02:00
await checkVideoFilesWereRemoved({ server: servers[0], video: video1Server2, onlyVideoFiles: true })
2020-01-10 10:11:28 +01:00
})
after(async function () {
await cleanupTests(servers)
})
})
describe('Test expiration', function () {
const strategy = 'recently-added'
2021-07-16 09:47:51 +02:00
async function checkContains (servers: PeerTubeServer[], str: string) {
for (const server of servers) {
2021-07-22 14:28:03 +02:00
const video = await server.videos.get({ id: video1Server2.uuid })
for (const f of video.files) {
expect(f.magnetUri).to.contain(str)
}
}
}
2021-07-16 09:47:51 +02:00
async function checkNotContains (servers: PeerTubeServer[], str: string) {
for (const server of servers) {
2021-07-22 14:28:03 +02:00
const video = await server.videos.get({ id: video1Server2.uuid })
for (const f of video.files) {
expect(f.magnetUri).to.not.contain(str)
}
}
}
before(async function () {
this.timeout(120000)
await createServers(strategy, { min_lifetime: '7 seconds', min_views: 0 })
await enableRedundancyOnServer1()
})
it('Should still have 2 webseeds after 10 seconds', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
await wait(10000)
try {
await checkContains(servers, 'http%3A%2F%2Flocalhost%3A' + servers[0].port)
} catch {
// Maybe a server deleted a redundancy in the scheduler
await wait(2000)
await checkContains(servers, 'http%3A%2F%2Flocalhost%3A' + servers[0].port)
}
})
it('Should stop server 1 and expire video redundancy', async function () {
2019-01-29 08:37:25 +01:00
this.timeout(80000)
2021-07-09 15:37:43 +02:00
await killallServers([ servers[0] ])
2018-11-15 11:20:23 +01:00
await wait(15000)
await checkNotContains([ servers[1], servers[2] ], 'http%3A%2F%2Flocalhost%3A' + servers[0].port)
})
2019-04-24 15:10:37 +02:00
after(async function () {
await cleanupTests(servers)
})
})
describe('Test file replacement', function () {
let video2Server2UUID: string
const strategy = 'recently-added'
before(async function () {
this.timeout(120000)
await createServers(strategy, { min_lifetime: '7 seconds', min_views: 0 })
await enableRedundancyOnServer1()
await waitJobs(servers)
2021-07-16 09:04:35 +02:00
await servers[0].servers.waitUntilLog('Duplicated ', 5)
await waitJobs(servers)
2021-07-22 14:28:03 +02:00
await check2Webseeds()
await check1PlaylistRedundancies()
2021-02-01 11:18:50 +01:00
await checkStatsWith1Redundancy(strategy)
2021-07-16 09:04:35 +02:00
const { uuid } = await servers[1].videos.upload({ attributes: { name: 'video 2 server 2', privacy: VideoPrivacy.PRIVATE } })
2021-07-15 10:02:54 +02:00
video2Server2UUID = uuid
2021-02-01 16:05:22 +01:00
// Wait transcoding before federation
await waitJobs(servers)
2021-07-16 09:04:35 +02:00
await servers[1].videos.update({ id: video2Server2UUID, attributes: { privacy: VideoPrivacy.PUBLIC } })
})
2018-11-15 11:20:23 +01:00
it('Should cache video 2 webseeds on the first video', async function () {
this.timeout(120000)
await waitJobs(servers)
2018-11-15 11:20:23 +01:00
let checked = false
2018-11-15 11:20:23 +01:00
while (checked === false) {
await wait(1000)
try {
2021-07-22 14:28:03 +02:00
await check1WebSeed()
await check0PlaylistRedundancies()
2021-02-01 16:05:22 +01:00
2019-01-29 08:37:25 +01:00
await check2Webseeds(video2Server2UUID)
await check1PlaylistRedundancies(video2Server2UUID)
2018-11-15 11:20:23 +01:00
checked = true
} catch {
checked = false
}
}
})
2019-01-29 08:37:25 +01:00
it('Should disable strategy and remove redundancies', async function () {
this.timeout(80000)
await waitJobs(servers)
2021-07-09 15:37:43 +02:00
await killallServers([ servers[0] ])
2021-07-16 09:47:51 +02:00
await servers[0].run({
2019-01-29 08:37:25 +01:00
redundancy: {
videos: {
check_interval: '1 second',
strategies: []
}
}
})
await waitJobs(servers)
2021-07-22 14:28:03 +02:00
await checkVideoFilesWereRemoved({ server: servers[0], video: video1Server2, onlyVideoFiles: true })
2019-01-29 08:37:25 +01:00
})
2019-04-24 15:10:37 +02:00
after(async function () {
await cleanupTests(servers)
})
})
})