mirror of
https://github.com/overleaf/overleaf.git
synced 2024-12-18 06:31:30 -05:00
b7c4f3333e
[server-pro] prepare hotfix release 4.1.4 GitOrigin-RevId: 2301f366f9b7f170a5801c5b74d10b9b7757973e
153 lines
5.2 KiB
Diff
153 lines
5.2 KiB
Diff
--- services/history-v1/api/controllers/projects.js
|
|
+++ services/history-v1/api/controllers/projects.js
|
|
@@ -194,18 +194,23 @@ async function getProjectBlob(req, res, next) {
|
|
const hash = req.swagger.params.hash.value
|
|
|
|
const blobStore = new BlobStore(projectId)
|
|
- let stream
|
|
+ logger.debug({ projectId, hash }, 'getProjectBlob started')
|
|
try {
|
|
- stream = await blobStore.getStream(hash)
|
|
- } catch (err) {
|
|
- if (err instanceof Blob.NotFoundError) {
|
|
- return render.notFound(res)
|
|
- } else {
|
|
- throw err
|
|
+ let stream
|
|
+ try {
|
|
+ stream = await blobStore.getStream(hash)
|
|
+ } catch (err) {
|
|
+ if (err instanceof Blob.NotFoundError) {
|
|
+ return render.notFound(res)
|
|
+ } else {
|
|
+ throw err
|
|
+ }
|
|
}
|
|
+ res.set('Content-Type', 'application/octet-stream')
|
|
+ await pipeline(stream, res)
|
|
+ } finally {
|
|
+ logger.debug({ projectId, hash }, 'getProjectBlob finished')
|
|
}
|
|
- res.set('Content-Type', 'application/octet-stream')
|
|
- await pipeline(stream, res)
|
|
}
|
|
|
|
async function getSnapshotAtVersion(projectId, version) {
|
|
--- services/history-v1/storage/lib/blob_store/index.js
|
|
+++ services/history-v1/storage/lib/blob_store/index.js
|
|
@@ -20,6 +20,7 @@ const projectKey = require('../project_key')
|
|
const streams = require('../streams')
|
|
const postgresBackend = require('./postgres')
|
|
const mongoBackend = require('./mongo')
|
|
+const logger = require('@overleaf/logger')
|
|
|
|
const GLOBAL_BLOBS = new Map()
|
|
|
|
@@ -34,9 +35,14 @@ function makeProjectKey(projectId, hash) {
|
|
async function uploadBlob(projectId, blob, stream) {
|
|
const bucket = config.get('blobStore.projectBucket')
|
|
const key = makeProjectKey(projectId, blob.getHash())
|
|
- await persistor.sendStream(bucket, key, stream, {
|
|
- contentType: 'application/octet-stream',
|
|
- })
|
|
+ logger.debug({ projectId, blob }, 'uploadBlob started')
|
|
+ try {
|
|
+ await persistor.sendStream(bucket, key, stream, {
|
|
+ contentType: 'application/octet-stream',
|
|
+ })
|
|
+ } finally {
|
|
+ logger.debug({ projectId, blob }, 'uploadBlob finished')
|
|
+ }
|
|
}
|
|
|
|
function getBlobLocation(projectId, hash) {
|
|
@@ -109,7 +115,12 @@ async function getStringLengthOfFile(byteLength, pathname) {
|
|
async function deleteBlobsInBucket(projectId) {
|
|
const bucket = config.get('blobStore.projectBucket')
|
|
const prefix = `${projectKey.format(projectId)}/`
|
|
- await persistor.deleteDirectory(bucket, prefix)
|
|
+ logger.debug({ projectId }, 'deleteBlobsInBucket started')
|
|
+ try {
|
|
+ await persistor.deleteDirectory(bucket, prefix)
|
|
+ } finally {
|
|
+ logger.debug({ projectId }, 'deleteBlobsInBucket finished')
|
|
+ }
|
|
}
|
|
|
|
async function loadGlobalBlobs() {
|
|
@@ -202,9 +213,15 @@ class BlobStore {
|
|
async getString(hash) {
|
|
assert.blobHash(hash, 'bad hash')
|
|
|
|
- const stream = await this.getStream(hash)
|
|
- const buffer = await streams.readStreamToBuffer(stream)
|
|
- return buffer.toString()
|
|
+ const projectId = this.projectId
|
|
+ logger.debug({ projectId, hash }, 'getString started')
|
|
+ try {
|
|
+ const stream = await this.getStream(hash)
|
|
+ const buffer = await streams.readStreamToBuffer(stream)
|
|
+ return buffer.toString()
|
|
+ } finally {
|
|
+ logger.debug({ projectId, hash }, 'getString finished')
|
|
+ }
|
|
}
|
|
|
|
/**
|
|
--- services/history-v1/storage/lib/history_store.js
|
|
+++ services/history-v1/storage/lib/history_store.js
|
|
@@ -8,6 +8,7 @@ const path = require('path')
|
|
|
|
const OError = require('@overleaf/o-error')
|
|
const objectPersistor = require('@overleaf/object-persistor')
|
|
+const logger = require('@overleaf/logger')
|
|
|
|
const assert = require('./assert')
|
|
const persistor = require('./persistor')
|
|
@@ -70,6 +71,7 @@ HistoryStore.prototype.loadRaw = function historyStoreLoadRaw(
|
|
|
|
const key = getKey(projectId, chunkId)
|
|
|
|
+ logger.debug({ projectId, chunkId }, 'loadRaw started')
|
|
return BPromise.resolve()
|
|
.then(() => persistor.getObjectStream(BUCKET, key))
|
|
.then(streams.gunzipStreamToBuffer)
|
|
@@ -80,6 +82,7 @@ HistoryStore.prototype.loadRaw = function historyStoreLoadRaw(
|
|
}
|
|
throw new HistoryStore.LoadError(projectId, chunkId).withCause(err)
|
|
})
|
|
+ .finally(() => logger.debug({ projectId, chunkId }, 'loadRaw finished'))
|
|
}
|
|
|
|
/**
|
|
@@ -102,6 +105,7 @@ HistoryStore.prototype.storeRaw = function historyStoreStoreRaw(
|
|
const key = getKey(projectId, chunkId)
|
|
const stream = streams.gzipStringToStream(JSON.stringify(rawHistory))
|
|
|
|
+ logger.debug({ projectId, chunkId }, 'storeRaw started')
|
|
return BPromise.resolve()
|
|
.then(() =>
|
|
persistor.sendStream(BUCKET, key, stream, {
|
|
@@ -112,6 +116,7 @@ HistoryStore.prototype.storeRaw = function historyStoreStoreRaw(
|
|
.catch(err => {
|
|
throw new HistoryStore.StoreError(projectId, chunkId).withCause(err)
|
|
})
|
|
+ .finally(() => logger.debug({ projectId, chunkId }, 'storeRaw finished'))
|
|
}
|
|
|
|
/**
|
|
@@ -121,12 +126,13 @@ HistoryStore.prototype.storeRaw = function historyStoreStoreRaw(
|
|
* @return {Promise}
|
|
*/
|
|
HistoryStore.prototype.deleteChunks = function historyDeleteChunks(chunks) {
|
|
+ logger.debug({ chunks }, 'deleteChunks started')
|
|
return BPromise.all(
|
|
chunks.map(chunk => {
|
|
const key = getKey(chunk.projectId, chunk.chunkId)
|
|
return persistor.deleteObject(BUCKET, key)
|
|
})
|
|
- )
|
|
+ ).finally(() => logger.debug({ chunks }, 'deleteChunks finished'))
|
|
}
|
|
|
|
module.exports = new HistoryStore()
|