mirror of
https://github.com/advplyr/audiobookshelf.git
synced 2024-12-29 01:58:49 +01:00
08250e266e
This patch implements [X-Accel](https://www.nginx.com/resources/wiki/start/topics/examples/x-accel/) redirect headers as an optional way for offloading static file delivery from Express to Nginx, which is far better optimized for static file delivery. This provides a really easy to configure way for getting a huge performance boost over delivering all files through Audiobookshelf. How it works ------------ The way this works is basically that Audiobookshelf gets an HTTP request for delivering a static file (let's say an audiobook). It will first check the user is authorized and then convert the API path to a local file path. Now, instead of reading and delivering the file, Audiobookshelf will return just the HTTP header with an additional `X-Accel-Redirect` pointing to the file location on the file syste. This header is picked up by Nginx which will then deliver the file. Configuration ------------- The configuration for this is very simple. You need to run Nginx as reverse proxy and it must have access to your Audiobookshelf data folder. You then configure Audiobookshelf to use X-Accel by setting `USE_X_ACCEL=/protected`. The path is the internal redirect path used by Nginx. In the Nginx configuration you then configure this location and map it to the storage area to serve like this: ``` location /protected/ { internal; alias /; } ``` That's all. Impact ------ I just did a very simple performance test, downloading a 1170620819 bytes large audiobook file from another machine on the same network like this, using `time -p` to measure how log the process took: ```sh URL='https://url to audiobook…' for i in `seq 1 50` do echo "$i" curl -s -o /dev/null "${URL}" done ``` This sequential test with 50 iterations and without x-accel resulted in: ``` real 413.42 user 197.11 sys 82.04 ``` That is an average download speed of about 1080 MBit/s. With X-Accel enabled, serving the files through Nginx, the same test yielded the following results: ``` real 200.37 user 86.95 sys 29.79 ``` That is an average download speed of about 2229 MBit/s, more than doubling the previous speed. I have also run the same test with 4 parallel processes and 25 downloads each. Without x-accel, that test resulted in: ``` real 364.89 user 273.09 sys 112.75 ``` That is an average speed of about 2448 MBit/s. With X-Accel enabled, the parallel test also shows a significant speedup: ``` real 167.19 user 195.62 sys 78.61 ``` That is an average speed of about 5342 MBit/s. While doing that, I also peaked at the system load which was a bit lower when using X-Accel. Even though the system was delivering far more data. But I just looked at the `load1` values and did not build a proper test for that. That means, I cant provide any definitive data. Supported Media --------------- The current implementation works for audio files and book covers. There are other media files which would benefit from this mechanism like feed covers or author pictures. But that's something for a future developer ;-)
168 lines
5.0 KiB
JavaScript
168 lines
5.0 KiB
JavaScript
const Path = require('path')
|
|
const fs = require('../libs/fsExtra')
|
|
const stream = require('stream')
|
|
const filePerms = require('../utils/filePerms')
|
|
const Logger = require('../Logger')
|
|
const { resizeImage } = require('../utils/ffmpegHelpers')
|
|
|
|
class CacheManager {
|
|
constructor() {
|
|
this.CachePath = Path.join(global.MetadataPath, 'cache')
|
|
this.CoverCachePath = Path.join(this.CachePath, 'covers')
|
|
this.ImageCachePath = Path.join(this.CachePath, 'images')
|
|
this.ItemCachePath = Path.join(this.CachePath, 'items')
|
|
}
|
|
|
|
async ensureCachePaths() { // Creates cache paths if necessary and sets owner and permissions
|
|
var pathsCreated = false
|
|
if (!(await fs.pathExists(this.CachePath))) {
|
|
await fs.mkdir(this.CachePath)
|
|
pathsCreated = true
|
|
}
|
|
|
|
if (!(await fs.pathExists(this.CoverCachePath))) {
|
|
await fs.mkdir(this.CoverCachePath)
|
|
pathsCreated = true
|
|
}
|
|
|
|
if (!(await fs.pathExists(this.ImageCachePath))) {
|
|
await fs.mkdir(this.ImageCachePath)
|
|
pathsCreated = true
|
|
}
|
|
|
|
if (!(await fs.pathExists(this.ItemCachePath))) {
|
|
await fs.mkdir(this.ItemCachePath)
|
|
pathsCreated = true
|
|
}
|
|
|
|
if (pathsCreated) {
|
|
await filePerms.setDefault(this.CachePath)
|
|
}
|
|
}
|
|
|
|
async handleCoverCache(res, libraryItem, options = {}) {
|
|
const format = options.format || 'webp'
|
|
const width = options.width || 400
|
|
const height = options.height || null
|
|
|
|
res.type(`image/${format}`)
|
|
|
|
const path = Path.join(this.CoverCachePath, `${libraryItem.id}_${width}${height ? `x${height}` : ''}`) + '.' + format
|
|
|
|
// Cache exists
|
|
if (await fs.pathExists(path)) {
|
|
if (global.XAccel) {
|
|
Logger.debug(`Use X-Accel to serve static file ${path}`)
|
|
return res.status(204).header({'X-Accel-Redirect': global.XAccel + path}).send()
|
|
}
|
|
|
|
const r = fs.createReadStream(path)
|
|
const ps = new stream.PassThrough()
|
|
stream.pipeline(r, ps, (err) => {
|
|
if (err) {
|
|
console.log(err)
|
|
return res.sendStatus(500)
|
|
}
|
|
})
|
|
return ps.pipe(res)
|
|
}
|
|
|
|
if (!libraryItem.media.coverPath || !await fs.pathExists(libraryItem.media.coverPath)) {
|
|
return res.sendStatus(500)
|
|
}
|
|
|
|
const writtenFile = await resizeImage(libraryItem.media.coverPath, path, width, height)
|
|
if (!writtenFile) return res.sendStatus(500)
|
|
|
|
// Set owner and permissions of cache image
|
|
await filePerms.setDefault(path)
|
|
|
|
if (global.XAccel) {
|
|
Logger.debug(`Use X-Accel to serve static file ${writtenFile}`)
|
|
return res.status(204).header({'X-Accel-Redirect': global.XAccel + writtenFile}).send()
|
|
}
|
|
|
|
var readStream = fs.createReadStream(writtenFile)
|
|
readStream.pipe(res)
|
|
}
|
|
|
|
purgeCoverCache(libraryItemId) {
|
|
return this.purgeEntityCache(libraryItemId, this.CoverCachePath)
|
|
}
|
|
|
|
purgeImageCache(entityId) {
|
|
return this.purgeEntityCache(entityId, this.ImageCachePath)
|
|
}
|
|
|
|
async purgeEntityCache(entityId, cachePath) {
|
|
return Promise.all((await fs.readdir(cachePath)).reduce((promises, file) => {
|
|
if (file.startsWith(entityId)) {
|
|
Logger.debug(`[CacheManager] Going to purge ${file}`);
|
|
promises.push(this.removeCache(Path.join(cachePath, file)))
|
|
}
|
|
return promises
|
|
}, []))
|
|
}
|
|
|
|
removeCache(path) {
|
|
if (!path) return false
|
|
return fs.pathExists(path).then((exists) => {
|
|
if (!exists) return false
|
|
return fs.unlink(path).then(() => true).catch((err) => {
|
|
Logger.error(`[CacheManager] Failed to remove cache "${path}"`, err)
|
|
return false
|
|
})
|
|
})
|
|
}
|
|
|
|
async purgeAll() {
|
|
if (await fs.pathExists(this.CachePath)) {
|
|
await fs.remove(this.CachePath).catch((error) => {
|
|
Logger.error(`[CacheManager] Failed to remove cache dir "${this.CachePath}"`, error)
|
|
})
|
|
}
|
|
await this.ensureCachePaths()
|
|
}
|
|
|
|
async purgeItems() {
|
|
if (await fs.pathExists(this.ItemCachePath)) {
|
|
await fs.remove(this.ItemCachePath).catch((error) => {
|
|
Logger.error(`[CacheManager] Failed to remove items cache dir "${this.ItemCachePath}"`, error)
|
|
})
|
|
}
|
|
await this.ensureCachePaths()
|
|
}
|
|
|
|
async handleAuthorCache(res, author, options = {}) {
|
|
const format = options.format || 'webp'
|
|
const width = options.width || 400
|
|
const height = options.height || null
|
|
|
|
res.type(`image/${format}`)
|
|
|
|
var path = Path.join(this.ImageCachePath, `${author.id}_${width}${height ? `x${height}` : ''}`) + '.' + format
|
|
|
|
// Cache exists
|
|
if (await fs.pathExists(path)) {
|
|
const r = fs.createReadStream(path)
|
|
const ps = new stream.PassThrough()
|
|
stream.pipeline(r, ps, (err) => {
|
|
if (err) {
|
|
console.log(err)
|
|
return res.sendStatus(500)
|
|
}
|
|
})
|
|
return ps.pipe(res)
|
|
}
|
|
|
|
let writtenFile = await resizeImage(author.imagePath, path, width, height)
|
|
if (!writtenFile) return res.sendStatus(500)
|
|
|
|
// Set owner and permissions of cache image
|
|
await filePerms.setDefault(path)
|
|
|
|
var readStream = fs.createReadStream(writtenFile)
|
|
readStream.pipe(res)
|
|
}
|
|
}
|
|
module.exports = CacheManager |