mirror of https://github.com/Chocobozzz/PeerTube
316 lines
8.7 KiB
TypeScript
316 lines
8.7 KiB
TypeScript
// Thanks: https://github.com/kwhitley/apicache
|
|
// We duplicated the library because it is unmaintened and prevent us to upgrade to recent NodeJS versions
|
|
|
|
import express from 'express'
|
|
import { OutgoingHttpHeaders } from 'http'
|
|
import { HttpStatusCodeType } from '@peertube/peertube-models'
|
|
import { isTestInstance } from '@peertube/peertube-node-utils'
|
|
import { parseDurationToMs } from '@server/helpers/core-utils.js'
|
|
import { logger } from '@server/helpers/logger.js'
|
|
import { Redis } from '@server/lib/redis.js'
|
|
import { asyncMiddleware } from '@server/middlewares/index.js'
|
|
|
|
export interface APICacheOptions {
|
|
headerBlacklist?: string[]
|
|
excludeStatus?: HttpStatusCodeType[]
|
|
}
|
|
|
|
interface CacheObject {
|
|
status: number
|
|
headers: OutgoingHttpHeaders
|
|
data: any
|
|
encoding: BufferEncoding
|
|
timestamp: number
|
|
}
|
|
|
|
export class ApiCache {
|
|
|
|
private readonly options: APICacheOptions
|
|
private readonly timers: { [ id: string ]: NodeJS.Timeout } = {}
|
|
|
|
private readonly index = {
|
|
groups: [] as string[],
|
|
all: [] as string[]
|
|
}
|
|
|
|
// Cache keys per group
|
|
private groups: { [groupIndex: string]: string[] } = {}
|
|
|
|
private readonly seed: number
|
|
|
|
constructor (options: APICacheOptions) {
|
|
this.seed = new Date().getTime()
|
|
|
|
this.options = {
|
|
headerBlacklist: [],
|
|
excludeStatus: [],
|
|
|
|
...options
|
|
}
|
|
}
|
|
|
|
buildMiddleware (strDuration: string) {
|
|
const duration = parseDurationToMs(strDuration)
|
|
|
|
return asyncMiddleware(
|
|
async (req: express.Request, res: express.Response, next: express.NextFunction) => {
|
|
const key = this.getCacheKey(req)
|
|
const redis = Redis.Instance.getClient()
|
|
|
|
if (!Redis.Instance.isConnected()) return this.makeResponseCacheable(res, next, key, duration)
|
|
|
|
try {
|
|
const obj = await redis.hgetall(key)
|
|
if (obj?.response) {
|
|
return this.sendCachedResponse(req, res, JSON.parse(obj.response), duration)
|
|
}
|
|
|
|
return this.makeResponseCacheable(res, next, key, duration)
|
|
} catch (err) {
|
|
return this.makeResponseCacheable(res, next, key, duration)
|
|
}
|
|
}
|
|
)
|
|
}
|
|
|
|
clearGroupSafe (group: string) {
|
|
const run = async () => {
|
|
const cacheKeys = this.groups[group]
|
|
if (!cacheKeys) return
|
|
|
|
for (const key of cacheKeys) {
|
|
try {
|
|
await this.clear(key)
|
|
} catch (err) {
|
|
logger.error('Cannot clear ' + key, { err })
|
|
}
|
|
}
|
|
|
|
delete this.groups[group]
|
|
}
|
|
|
|
void run()
|
|
}
|
|
|
|
private getCacheKey (req: express.Request) {
|
|
return Redis.Instance.getPrefix() + 'api-cache-' + this.seed + '-' + req.originalUrl
|
|
}
|
|
|
|
private shouldCacheResponse (response: express.Response) {
|
|
if (!response) return false
|
|
if (this.options.excludeStatus.includes(response.statusCode as HttpStatusCodeType)) return false
|
|
|
|
return true
|
|
}
|
|
|
|
private addIndexEntries (key: string, res: express.Response) {
|
|
this.index.all.unshift(key)
|
|
|
|
const groups = res.locals.apicacheGroups || []
|
|
|
|
for (const group of groups) {
|
|
if (!this.groups[group]) this.groups[group] = []
|
|
|
|
this.groups[group].push(key)
|
|
}
|
|
}
|
|
|
|
private filterBlacklistedHeaders (headers: OutgoingHttpHeaders) {
|
|
return Object.keys(headers)
|
|
.filter(key => !this.options.headerBlacklist.includes(key))
|
|
.reduce((acc, header) => {
|
|
acc[header] = headers[header]
|
|
|
|
return acc
|
|
}, {})
|
|
}
|
|
|
|
private createCacheObject (status: number, headers: OutgoingHttpHeaders, data: any, encoding: BufferEncoding) {
|
|
return {
|
|
status,
|
|
headers: this.filterBlacklistedHeaders(headers),
|
|
data,
|
|
encoding,
|
|
|
|
// Seconds since epoch, used to properly decrement max-age headers in cached responses.
|
|
timestamp: new Date().getTime() / 1000
|
|
} as CacheObject
|
|
}
|
|
|
|
private async cacheResponse (key: string, value: object, duration: number) {
|
|
const redis = Redis.Instance.getClient()
|
|
|
|
if (Redis.Instance.isConnected()) {
|
|
await Promise.all([
|
|
redis.hset(key, 'response', JSON.stringify(value)),
|
|
redis.hset(key, 'duration', duration + ''),
|
|
redis.expire(key, duration / 1000)
|
|
])
|
|
}
|
|
|
|
// add automatic cache clearing from duration, includes max limit on setTimeout
|
|
this.timers[key] = setTimeout(() => {
|
|
this.clear(key)
|
|
.catch(err => logger.error('Cannot clear Redis key %s.', key, { err }))
|
|
}, Math.min(duration, 2147483647))
|
|
}
|
|
|
|
private accumulateContent (res: express.Response, content: any) {
|
|
if (!content) return
|
|
|
|
if (typeof content === 'string') {
|
|
res.locals.apicache.content = (res.locals.apicache.content || '') + content
|
|
return
|
|
}
|
|
|
|
if (Buffer.isBuffer(content)) {
|
|
let oldContent = res.locals.apicache.content
|
|
|
|
if (typeof oldContent === 'string') {
|
|
oldContent = Buffer.from(oldContent)
|
|
}
|
|
|
|
if (!oldContent) {
|
|
oldContent = Buffer.alloc(0)
|
|
}
|
|
|
|
res.locals.apicache.content = Buffer.concat(
|
|
[ oldContent, content ],
|
|
oldContent.length + content.length
|
|
)
|
|
|
|
return
|
|
}
|
|
|
|
res.locals.apicache.content = content
|
|
}
|
|
|
|
private makeResponseCacheable (res: express.Response, next: express.NextFunction, key: string, duration: number) {
|
|
const self = this
|
|
|
|
res.locals.apicache = {
|
|
write: res.write.bind(res),
|
|
writeHead: res.writeHead.bind(res),
|
|
end: res.end.bind(res),
|
|
cacheable: true,
|
|
content: undefined,
|
|
headers: undefined
|
|
}
|
|
|
|
// Patch express
|
|
res.writeHead = function () {
|
|
if (self.shouldCacheResponse(res)) {
|
|
res.setHeader('cache-control', 'max-age=' + (duration / 1000).toFixed(0))
|
|
} else {
|
|
res.setHeader('cache-control', 'no-cache, no-store, must-revalidate')
|
|
}
|
|
|
|
res.locals.apicache.headers = Object.assign({}, res.getHeaders())
|
|
return res.locals.apicache.writeHead.apply(this, arguments as any)
|
|
}
|
|
|
|
res.write = function (chunk: any) {
|
|
self.accumulateContent(res, chunk)
|
|
return res.locals.apicache.write.apply(this, arguments as any)
|
|
}
|
|
|
|
res.end = function (content: any, encoding: BufferEncoding) {
|
|
if (self.shouldCacheResponse(res)) {
|
|
self.accumulateContent(res, content)
|
|
|
|
if (res.locals.apicache.cacheable && res.locals.apicache.content) {
|
|
self.addIndexEntries(key, res)
|
|
|
|
const headers = res.locals.apicache.headers || res.getHeaders()
|
|
const cacheObject = self.createCacheObject(
|
|
res.statusCode,
|
|
headers,
|
|
res.locals.apicache.content,
|
|
encoding
|
|
)
|
|
self.cacheResponse(key, cacheObject, duration)
|
|
.catch(err => logger.error('Cannot cache response', { err }))
|
|
}
|
|
}
|
|
|
|
res.locals.apicache.end.apply(this, arguments as any)
|
|
} as any
|
|
|
|
next()
|
|
}
|
|
|
|
private sendCachedResponse (request: express.Request, response: express.Response, cacheObject: CacheObject, duration: number) {
|
|
const headers = response.getHeaders()
|
|
|
|
if (isTestInstance()) {
|
|
Object.assign(headers, {
|
|
'x-api-cache-cached': 'true'
|
|
})
|
|
}
|
|
|
|
Object.assign(headers, this.filterBlacklistedHeaders(cacheObject.headers || {}), {
|
|
// Set properly decremented max-age header
|
|
// This ensures that max-age is in sync with the cache expiration
|
|
'cache-control':
|
|
'max-age=' +
|
|
Math.max(
|
|
0,
|
|
(duration / 1000 - (new Date().getTime() / 1000 - cacheObject.timestamp))
|
|
).toFixed(0)
|
|
})
|
|
|
|
// unstringify buffers
|
|
let data = cacheObject.data
|
|
if (data && data.type === 'Buffer') {
|
|
data = typeof data.data === 'number'
|
|
? Buffer.alloc(data.data)
|
|
: Buffer.from(data.data)
|
|
}
|
|
|
|
// Test Etag against If-None-Match for 304
|
|
const cachedEtag = cacheObject.headers.etag
|
|
const requestEtag = request.headers['if-none-match']
|
|
|
|
if (requestEtag && cachedEtag === requestEtag) {
|
|
response.writeHead(304, headers)
|
|
return response.end()
|
|
}
|
|
|
|
response.writeHead(cacheObject.status || 200, headers)
|
|
|
|
return response.end(data, cacheObject.encoding)
|
|
}
|
|
|
|
private async clear (target: string) {
|
|
const redis = Redis.Instance.getClient()
|
|
|
|
if (target) {
|
|
clearTimeout(this.timers[target])
|
|
delete this.timers[target]
|
|
|
|
try {
|
|
await redis.del(target)
|
|
} catch (err) {
|
|
logger.error('Cannot delete %s in redis cache.', target, { err })
|
|
}
|
|
|
|
this.index.all = this.index.all.filter(key => key !== target)
|
|
} else {
|
|
for (const key of this.index.all) {
|
|
clearTimeout(this.timers[key])
|
|
delete this.timers[key]
|
|
|
|
try {
|
|
await redis.del(key)
|
|
} catch (err) {
|
|
logger.error('Cannot delete %s in redis cache.', key, { err })
|
|
}
|
|
}
|
|
|
|
this.index.all = []
|
|
}
|
|
|
|
return this.index
|
|
}
|
|
}
|