mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-21 20:47:08 -05:00
287 lines
7.3 KiB
JavaScript
287 lines
7.3 KiB
JavaScript
const settings = require('settings-sharelatex')
|
|
const fs = require('fs')
|
|
const { promisify } = require('util')
|
|
const Stream = require('stream')
|
|
const { Storage } = require('@google-cloud/storage')
|
|
const { callbackify } = require('util')
|
|
const { WriteError, ReadError, NotFoundError } = require('./Errors')
|
|
const asyncPool = require('tiny-async-pool')
|
|
const PersistorHelper = require('./PersistorHelper')
|
|
|
|
const pipeline = promisify(Stream.pipeline)
|
|
|
|
// endpoint settings will be null by default except for tests
|
|
// that's OK - GCS uses the locally-configured service account by default
|
|
const storage = new Storage(settings.filestore.gcs.endpoint)
|
|
// workaround for broken uploads with custom endpoints:
|
|
// https://github.com/googleapis/nodejs-storage/issues/898
|
|
if (
|
|
settings.filestore.gcs.endpoint &&
|
|
settings.filestore.gcs.endpoint.apiEndpoint
|
|
) {
|
|
storage.interceptors.push({
|
|
request: function(reqOpts) {
|
|
const url = new URL(reqOpts.uri)
|
|
url.host = settings.filestore.gcs.endpoint.apiEndpoint
|
|
if (settings.filestore.gcs.endpoint.apiScheme) {
|
|
url.protocol = settings.filestore.gcs.endpoint.apiScheme
|
|
}
|
|
reqOpts.uri = url.toString()
|
|
return reqOpts
|
|
}
|
|
})
|
|
}
|
|
|
|
const GcsPersistor = {
|
|
sendFile: callbackify(sendFile),
|
|
sendStream: callbackify(sendStream),
|
|
getFileStream: callbackify(getFileStream),
|
|
getFileMd5Hash: callbackify(getFileMd5Hash),
|
|
deleteDirectory: callbackify(deleteDirectory),
|
|
getFileSize: callbackify(getFileSize),
|
|
deleteFile: callbackify(deleteFile),
|
|
copyFile: callbackify(copyFile),
|
|
checkIfFileExists: callbackify(checkIfFileExists),
|
|
directorySize: callbackify(directorySize),
|
|
promises: {
|
|
sendFile,
|
|
sendStream,
|
|
getFileStream,
|
|
getFileMd5Hash,
|
|
deleteDirectory,
|
|
getFileSize,
|
|
deleteFile,
|
|
copyFile,
|
|
checkIfFileExists,
|
|
directorySize
|
|
}
|
|
}
|
|
|
|
module.exports = GcsPersistor
|
|
|
|
async function sendFile(bucketName, key, fsPath) {
|
|
return sendStream(bucketName, key, fs.createReadStream(fsPath))
|
|
}
|
|
|
|
async function sendStream(bucketName, key, readStream, sourceMd5) {
|
|
try {
|
|
// egress from us to gcs
|
|
const observeOptions = { metric: 'gcs.egress' }
|
|
|
|
if (!sourceMd5) {
|
|
// if there is no supplied md5 hash, we calculate the hash as the data passes through
|
|
observeOptions.hash = 'md5'
|
|
}
|
|
|
|
const observer = new PersistorHelper.ObserverStream(observeOptions)
|
|
|
|
const writeOptions = {
|
|
// disabling of resumable uploads is recommended by Google:
|
|
resumable: false
|
|
}
|
|
|
|
if (sourceMd5) {
|
|
writeOptions.validation = 'md5'
|
|
writeOptions.metadata = {
|
|
md5Hash: PersistorHelper.hexToBase64(sourceMd5)
|
|
}
|
|
}
|
|
|
|
const uploadStream = storage
|
|
.bucket(bucketName)
|
|
.file(key)
|
|
.createWriteStream(writeOptions)
|
|
|
|
await pipeline(readStream, observer, uploadStream)
|
|
|
|
// if we didn't have an md5 hash, we should compare our computed one with Google's
|
|
// as we couldn't tell GCS about it beforehand
|
|
if (!sourceMd5) {
|
|
sourceMd5 = observer.getHash()
|
|
// throws on mismatch
|
|
await PersistorHelper.verifyMd5(GcsPersistor, bucketName, key, sourceMd5)
|
|
}
|
|
} catch (err) {
|
|
throw PersistorHelper.wrapError(
|
|
err,
|
|
'upload to GCS failed',
|
|
{ bucketName, key },
|
|
WriteError
|
|
)
|
|
}
|
|
}
|
|
|
|
async function getFileStream(bucketName, key, _opts = {}) {
|
|
const opts = Object.assign({}, _opts)
|
|
if (opts.end) {
|
|
// S3 (and http range headers) treat 'end' as inclusive, so increase this by 1
|
|
opts.end++
|
|
}
|
|
const stream = storage
|
|
.bucket(bucketName)
|
|
.file(key)
|
|
.createReadStream(opts)
|
|
|
|
// ingress to us from gcs
|
|
const observer = new PersistorHelper.ObserverStream({
|
|
metric: 'gcs.ingress'
|
|
})
|
|
|
|
try {
|
|
// wait for the pipeline to be ready, to catch non-200s
|
|
await PersistorHelper.getReadyPipeline(stream, observer)
|
|
return observer
|
|
} catch (err) {
|
|
throw PersistorHelper.wrapError(
|
|
err,
|
|
'error reading file from GCS',
|
|
{ bucketName, key, opts },
|
|
ReadError
|
|
)
|
|
}
|
|
}
|
|
|
|
async function getFileSize(bucketName, key) {
|
|
try {
|
|
const [metadata] = await storage
|
|
.bucket(bucketName)
|
|
.file(key)
|
|
.getMetadata()
|
|
return metadata.size
|
|
} catch (err) {
|
|
throw PersistorHelper.wrapError(
|
|
err,
|
|
'error getting size of GCS object',
|
|
{ bucketName, key },
|
|
ReadError
|
|
)
|
|
}
|
|
}
|
|
|
|
async function getFileMd5Hash(bucketName, key) {
|
|
try {
|
|
const [metadata] = await storage
|
|
.bucket(bucketName)
|
|
.file(key)
|
|
.getMetadata()
|
|
return PersistorHelper.base64ToHex(metadata.md5Hash)
|
|
} catch (err) {
|
|
throw PersistorHelper.wrapError(
|
|
err,
|
|
'error getting hash of GCS object',
|
|
{ bucketName, key },
|
|
ReadError
|
|
)
|
|
}
|
|
}
|
|
|
|
async function deleteFile(bucketName, key) {
|
|
try {
|
|
const file = storage.bucket(bucketName).file(key)
|
|
|
|
if (settings.filestore.gcs.deletedBucketSuffix) {
|
|
await file.copy(
|
|
storage
|
|
.bucket(`${bucketName}${settings.filestore.gcs.deletedBucketSuffix}`)
|
|
.file(`${key}-${new Date().toISOString()}`)
|
|
)
|
|
}
|
|
if (settings.filestore.gcs.unlockBeforeDelete) {
|
|
await file.setMetadata({ eventBasedHold: false })
|
|
}
|
|
await file.delete()
|
|
} catch (err) {
|
|
const error = PersistorHelper.wrapError(
|
|
err,
|
|
'error deleting GCS object',
|
|
{ bucketName, key },
|
|
WriteError
|
|
)
|
|
if (!(error instanceof NotFoundError)) {
|
|
throw error
|
|
}
|
|
}
|
|
}
|
|
|
|
async function deleteDirectory(bucketName, key) {
|
|
try {
|
|
const [files] = await storage
|
|
.bucket(bucketName)
|
|
.getFiles({ directory: key })
|
|
|
|
await asyncPool(
|
|
settings.filestore.gcs.deleteConcurrency,
|
|
files,
|
|
async file => {
|
|
await deleteFile(bucketName, file.name)
|
|
}
|
|
)
|
|
} catch (err) {
|
|
const error = PersistorHelper.wrapError(
|
|
err,
|
|
'failed to delete directory in GCS',
|
|
{ bucketName, key },
|
|
WriteError
|
|
)
|
|
if (error instanceof NotFoundError) {
|
|
return
|
|
}
|
|
throw error
|
|
}
|
|
}
|
|
|
|
async function directorySize(bucketName, key) {
|
|
let files
|
|
|
|
try {
|
|
const [response] = await storage
|
|
.bucket(bucketName)
|
|
.getFiles({ directory: key })
|
|
files = response
|
|
} catch (err) {
|
|
throw PersistorHelper.wrapError(
|
|
err,
|
|
'failed to list objects in GCS',
|
|
{ bucketName, key },
|
|
ReadError
|
|
)
|
|
}
|
|
|
|
return files.reduce((acc, file) => Number(file.metadata.size) + acc, 0)
|
|
}
|
|
|
|
async function checkIfFileExists(bucketName, key) {
|
|
try {
|
|
const [response] = await storage
|
|
.bucket(bucketName)
|
|
.file(key)
|
|
.exists()
|
|
return response
|
|
} catch (err) {
|
|
throw PersistorHelper.wrapError(
|
|
err,
|
|
'error checking if file exists in GCS',
|
|
{ bucketName, key },
|
|
ReadError
|
|
)
|
|
}
|
|
}
|
|
|
|
async function copyFile(bucketName, sourceKey, destKey) {
|
|
try {
|
|
const src = storage.bucket(bucketName).file(sourceKey)
|
|
const dest = storage.bucket(bucketName).file(destKey)
|
|
await src.copy(dest)
|
|
} catch (err) {
|
|
// fake-gcs-server has a bug that returns an invalid response when the file does not exist
|
|
if (err.message === 'Cannot parse response as JSON: not found\n') {
|
|
err.code = 404
|
|
}
|
|
throw PersistorHelper.wrapError(
|
|
err,
|
|
'failed to copy file in GCS',
|
|
{ bucketName, sourceKey, destKey },
|
|
WriteError
|
|
)
|
|
}
|
|
}
|