mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-21 20:47:08 -05:00
7f48c67512
* Add `unicorn/prefer-node-protocol` * Fix `unicorn/prefer-node-protocol` ESLint errors * Run `npm run format:fix` * Add sandboxed-module sourceTransformers in mocha setups Fix `no such file or directory, open 'node:fs'` in `sandboxed-module` * Remove `node:` in the SandboxedModule requires * Fix new linting errors with `node:` GitOrigin-RevId: 68f6e31e2191fcff4cb8058dd0a6914c14f59926
796 lines
25 KiB
JavaScript
796 lines
25 KiB
JavaScript
const Settings = require('@overleaf/settings')
|
|
const rclient = require('@overleaf/redis-wrapper').createClient(
|
|
Settings.redis.documentupdater
|
|
)
|
|
const logger = require('@overleaf/logger')
|
|
const OError = require('@overleaf/o-error')
|
|
const { promisifyAll } = require('@overleaf/promise-utils')
|
|
const metrics = require('./Metrics')
|
|
const Errors = require('./Errors')
|
|
const crypto = require('node:crypto')
|
|
const async = require('async')
|
|
const { docIsTooLarge } = require('./Limits')
|
|
|
|
// Sometimes Redis calls take an unexpectedly long time. We have to be
|
|
// quick with Redis calls because we're holding a lock that expires
|
|
// after 30 seconds. We can't let any errors in the rest of the stack
|
|
// hold us up, and need to bail out quickly if there is a problem.
|
|
const MAX_REDIS_REQUEST_LENGTH = 5000 // 5 seconds
|
|
const PROJECT_BLOCK_TTL_SECS = 30
|
|
|
|
// Make times easy to read
|
|
const minutes = 60 // seconds for Redis expire
|
|
|
|
const logHashReadErrors = Settings.documentupdater?.logHashErrors?.read
|
|
|
|
const MEGABYTES = 1024 * 1024
|
|
const MAX_RANGES_SIZE = 3 * MEGABYTES
|
|
|
|
const keys = Settings.redis.documentupdater.key_schema
|
|
|
|
const RedisManager = {
|
|
rclient,
|
|
|
|
putDocInMemory(
|
|
projectId,
|
|
docId,
|
|
docLines,
|
|
version,
|
|
ranges,
|
|
resolvedCommentIds,
|
|
pathname,
|
|
projectHistoryId,
|
|
historyRangesSupport,
|
|
_callback
|
|
) {
|
|
const timer = new metrics.Timer('redis.put-doc')
|
|
const callback = error => {
|
|
timer.done()
|
|
_callback(error)
|
|
}
|
|
const docLinesArray = docLines
|
|
docLines = JSON.stringify(docLines)
|
|
if (docLines.indexOf('\u0000') !== -1) {
|
|
const error = new Error('null bytes found in doc lines')
|
|
// this check was added to catch memory corruption in JSON.stringify.
|
|
// It sometimes returned null bytes at the end of the string.
|
|
logger.error({ err: error, docId, docLines }, error.message)
|
|
return callback(error)
|
|
}
|
|
// Do an optimised size check on the docLines using the serialised
|
|
// length as an upper bound
|
|
const sizeBound = docLines.length
|
|
if (docIsTooLarge(sizeBound, docLinesArray, Settings.max_doc_length)) {
|
|
const docSize = docLines.length
|
|
const err = new Error('blocking doc insert into redis: doc is too large')
|
|
logger.error({ projectId, docId, err, docSize }, err.message)
|
|
return callback(err)
|
|
}
|
|
const docHash = RedisManager._computeHash(docLines)
|
|
// record bytes sent to redis
|
|
metrics.summary('redis.docLines', docLines.length, { status: 'set' })
|
|
logger.debug(
|
|
{ projectId, docId, version, docHash, pathname, projectHistoryId },
|
|
'putting doc in redis'
|
|
)
|
|
RedisManager._serializeRanges(ranges, (error, ranges) => {
|
|
if (error) {
|
|
logger.error({ err: error, docId, projectId }, error.message)
|
|
return callback(error)
|
|
}
|
|
|
|
// update docsInProject set before writing doc contents
|
|
const multi = rclient.multi()
|
|
multi.exists(keys.projectBlock({ project_id: projectId }))
|
|
multi.sadd(keys.docsInProject({ project_id: projectId }), docId)
|
|
multi.exec((err, reply) => {
|
|
if (err) {
|
|
return callback(err)
|
|
}
|
|
const projectBlocked = reply[0] === 1
|
|
if (projectBlocked) {
|
|
// We don't clean up the spurious docId added in the docsInProject
|
|
// set. There is a risk that the docId was successfully added by a
|
|
// concurrent process. This set is used when unloading projects. An
|
|
// extra docId will not prevent the project from being uploaded, but
|
|
// a missing docId means that the doc might stay in Redis forever.
|
|
return callback(
|
|
new OError('Project blocked from loading docs', { projectId })
|
|
)
|
|
}
|
|
|
|
RedisManager.setHistoryRangesSupportFlag(
|
|
docId,
|
|
historyRangesSupport,
|
|
err => {
|
|
if (err) {
|
|
return callback(err)
|
|
}
|
|
|
|
if (!pathname) {
|
|
metrics.inc('pathname', 1, {
|
|
path: 'RedisManager.setDoc',
|
|
status: pathname === '' ? 'zero-length' : 'undefined',
|
|
})
|
|
}
|
|
|
|
// Make sure that this MULTI operation only operates on doc
|
|
// specific keys, i.e. keys that have the doc id in curly braces.
|
|
// The curly braces identify a hash key for Redis and ensures that
|
|
// the MULTI's operations are all done on the same node in a
|
|
// cluster environment.
|
|
const multi = rclient.multi()
|
|
multi.mset({
|
|
[keys.docLines({ doc_id: docId })]: docLines,
|
|
[keys.projectKey({ doc_id: docId })]: projectId,
|
|
[keys.docVersion({ doc_id: docId })]: version,
|
|
[keys.docHash({ doc_id: docId })]: docHash,
|
|
[keys.ranges({ doc_id: docId })]: ranges,
|
|
[keys.pathname({ doc_id: docId })]: pathname,
|
|
[keys.projectHistoryId({ doc_id: docId })]: projectHistoryId,
|
|
})
|
|
if (historyRangesSupport) {
|
|
multi.del(keys.resolvedCommentIds({ doc_id: docId }))
|
|
if (resolvedCommentIds.length > 0) {
|
|
multi.sadd(
|
|
keys.resolvedCommentIds({ doc_id: docId }),
|
|
...resolvedCommentIds
|
|
)
|
|
}
|
|
}
|
|
multi.exec(err => {
|
|
if (err) {
|
|
callback(
|
|
OError.tag(err, 'failed to write doc to Redis in MULTI', {
|
|
previousErrors: err.previousErrors.map(e => ({
|
|
name: e.name,
|
|
message: e.message,
|
|
command: e.command,
|
|
})),
|
|
})
|
|
)
|
|
} else {
|
|
callback()
|
|
}
|
|
})
|
|
}
|
|
)
|
|
})
|
|
})
|
|
},
|
|
|
|
removeDocFromMemory(projectId, docId, _callback) {
|
|
logger.debug({ projectId, docId }, 'removing doc from redis')
|
|
const callback = err => {
|
|
if (err) {
|
|
logger.err({ projectId, docId, err }, 'error removing doc from redis')
|
|
_callback(err)
|
|
} else {
|
|
logger.debug({ projectId, docId }, 'removed doc from redis')
|
|
_callback()
|
|
}
|
|
}
|
|
|
|
// Make sure that this MULTI operation only operates on doc
|
|
// specific keys, i.e. keys that have the doc id in curly braces.
|
|
// The curly braces identify a hash key for Redis and ensures that
|
|
// the MULTI's operations are all done on the same node in a
|
|
// cluster environment.
|
|
let multi = rclient.multi()
|
|
multi.strlen(keys.docLines({ doc_id: docId }))
|
|
multi.del(
|
|
keys.docLines({ doc_id: docId }),
|
|
keys.projectKey({ doc_id: docId }),
|
|
keys.docVersion({ doc_id: docId }),
|
|
keys.docHash({ doc_id: docId }),
|
|
keys.ranges({ doc_id: docId }),
|
|
keys.pathname({ doc_id: docId }),
|
|
keys.projectHistoryId({ doc_id: docId }),
|
|
keys.unflushedTime({ doc_id: docId }),
|
|
keys.lastUpdatedAt({ doc_id: docId }),
|
|
keys.lastUpdatedBy({ doc_id: docId }),
|
|
keys.resolvedCommentIds({ doc_id: docId })
|
|
)
|
|
multi.exec((error, response) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
const length = response?.[0]
|
|
if (length > 0) {
|
|
// record bytes freed in redis
|
|
metrics.summary('redis.docLines', length, { status: 'del' })
|
|
}
|
|
|
|
// Make sure that this MULTI operation only operates on project
|
|
// specific keys, i.e. keys that have the project id in curly braces.
|
|
// The curly braces identify a hash key for Redis and ensures that
|
|
// the MULTI's operations are all done on the same node in a
|
|
// cluster environment.
|
|
multi = rclient.multi()
|
|
multi.srem(keys.docsInProject({ project_id: projectId }), docId)
|
|
multi.del(keys.projectState({ project_id: projectId }))
|
|
multi.exec(err => {
|
|
if (err) {
|
|
return callback(err)
|
|
}
|
|
rclient.srem(keys.historyRangesSupport(), docId, callback)
|
|
})
|
|
})
|
|
},
|
|
|
|
checkOrSetProjectState(projectId, newState, callback) {
|
|
// Make sure that this MULTI operation only operates on project
|
|
// specific keys, i.e. keys that have the project id in curly braces.
|
|
// The curly braces identify a hash key for Redis and ensures that
|
|
// the MULTI's operations are all done on the same node in a
|
|
// cluster environment.
|
|
const multi = rclient.multi()
|
|
multi.getset(keys.projectState({ project_id: projectId }), newState)
|
|
multi.expire(keys.projectState({ project_id: projectId }), 30 * minutes)
|
|
multi.exec((error, response) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
logger.debug(
|
|
{ projectId, newState, oldState: response[0] },
|
|
'checking project state'
|
|
)
|
|
callback(null, response[0] !== newState)
|
|
})
|
|
},
|
|
|
|
clearProjectState(projectId, callback) {
|
|
rclient.del(keys.projectState({ project_id: projectId }), callback)
|
|
},
|
|
|
|
getDoc(projectId, docId, callback) {
|
|
const timer = new metrics.Timer('redis.get-doc')
|
|
const collectKeys = [
|
|
keys.docLines({ doc_id: docId }),
|
|
keys.docVersion({ doc_id: docId }),
|
|
keys.docHash({ doc_id: docId }),
|
|
keys.projectKey({ doc_id: docId }),
|
|
keys.ranges({ doc_id: docId }),
|
|
keys.pathname({ doc_id: docId }),
|
|
keys.projectHistoryId({ doc_id: docId }),
|
|
keys.unflushedTime({ doc_id: docId }),
|
|
keys.lastUpdatedAt({ doc_id: docId }),
|
|
keys.lastUpdatedBy({ doc_id: docId }),
|
|
]
|
|
rclient.mget(...collectKeys, (error, result) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
let [
|
|
docLines,
|
|
version,
|
|
storedHash,
|
|
docProjectId,
|
|
ranges,
|
|
pathname,
|
|
projectHistoryId,
|
|
unflushedTime,
|
|
lastUpdatedAt,
|
|
lastUpdatedBy,
|
|
] = result
|
|
rclient.sismember(keys.historyRangesSupport(), docId, (error, result) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
rclient.smembers(
|
|
keys.resolvedCommentIds({ doc_id: docId }),
|
|
(error, resolvedCommentIds) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
|
|
const historyRangesSupport = result === 1
|
|
|
|
const timeSpan = timer.done()
|
|
// check if request took too long and bail out. only do this for
|
|
// get, because it is the first call in each update, so if this
|
|
// passes we'll assume others have a reasonable chance to succeed.
|
|
if (timeSpan > MAX_REDIS_REQUEST_LENGTH) {
|
|
error = new Error('redis getDoc exceeded timeout')
|
|
return callback(error)
|
|
}
|
|
// record bytes loaded from redis
|
|
if (docLines != null) {
|
|
metrics.summary('redis.docLines', docLines.length, {
|
|
status: 'get',
|
|
})
|
|
}
|
|
// check sha1 hash value if present
|
|
if (docLines != null && storedHash != null) {
|
|
const computedHash = RedisManager._computeHash(docLines)
|
|
if (logHashReadErrors && computedHash !== storedHash) {
|
|
logger.error(
|
|
{
|
|
projectId,
|
|
docId,
|
|
docProjectId,
|
|
computedHash,
|
|
storedHash,
|
|
docLines,
|
|
},
|
|
'hash mismatch on retrieved document'
|
|
)
|
|
}
|
|
}
|
|
|
|
try {
|
|
docLines = JSON.parse(docLines)
|
|
ranges = RedisManager._deserializeRanges(ranges)
|
|
} catch (e) {
|
|
return callback(e)
|
|
}
|
|
|
|
version = parseInt(version || 0, 10)
|
|
// check doc is in requested project
|
|
if (docProjectId != null && docProjectId !== projectId) {
|
|
logger.error(
|
|
{ projectId, docId, docProjectId },
|
|
'doc not in project'
|
|
)
|
|
return callback(new Errors.NotFoundError('document not found'))
|
|
}
|
|
|
|
if (docLines && version && !pathname) {
|
|
metrics.inc('pathname', 1, {
|
|
path: 'RedisManager.getDoc',
|
|
status: pathname === '' ? 'zero-length' : 'undefined',
|
|
})
|
|
}
|
|
|
|
callback(
|
|
null,
|
|
docLines,
|
|
version,
|
|
ranges,
|
|
pathname,
|
|
projectHistoryId,
|
|
unflushedTime,
|
|
lastUpdatedAt,
|
|
lastUpdatedBy,
|
|
historyRangesSupport,
|
|
resolvedCommentIds
|
|
)
|
|
}
|
|
)
|
|
})
|
|
})
|
|
},
|
|
|
|
getDocVersion(docId, callback) {
|
|
rclient.mget(keys.docVersion({ doc_id: docId }), (error, result) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
let [version] = result || []
|
|
version = parseInt(version, 10)
|
|
callback(null, version)
|
|
})
|
|
},
|
|
|
|
getDocLines(docId, callback) {
|
|
rclient.get(keys.docLines({ doc_id: docId }), (error, docLines) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
callback(null, docLines)
|
|
})
|
|
},
|
|
|
|
getPreviousDocOps(docId, start, end, callback) {
|
|
const timer = new metrics.Timer('redis.get-prev-docops')
|
|
rclient.llen(keys.docOps({ doc_id: docId }), (error, length) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
rclient.get(keys.docVersion({ doc_id: docId }), (error, version) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
version = parseInt(version, 10)
|
|
const firstVersionInRedis = version - length
|
|
|
|
if (start < firstVersionInRedis || end > version) {
|
|
error = new Errors.OpRangeNotAvailableError(
|
|
'doc ops range is not loaded in redis',
|
|
{ firstVersionInRedis, version, ttlInS: RedisManager.DOC_OPS_TTL }
|
|
)
|
|
logger.debug(
|
|
{ err: error, docId, length, version, start, end },
|
|
'doc ops range is not loaded in redis'
|
|
)
|
|
return callback(error)
|
|
}
|
|
|
|
start = start - firstVersionInRedis
|
|
if (end > -1) {
|
|
end = end - firstVersionInRedis
|
|
}
|
|
|
|
if (isNaN(start) || isNaN(end)) {
|
|
error = new Error('inconsistent version or lengths')
|
|
logger.error(
|
|
{ err: error, docId, length, version, start, end },
|
|
'inconsistent version or length'
|
|
)
|
|
return callback(error)
|
|
}
|
|
|
|
rclient.lrange(
|
|
keys.docOps({ doc_id: docId }),
|
|
start,
|
|
end,
|
|
(error, jsonOps) => {
|
|
let ops
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
try {
|
|
ops = jsonOps.map(jsonOp => JSON.parse(jsonOp))
|
|
} catch (e) {
|
|
return callback(e)
|
|
}
|
|
const timeSpan = timer.done()
|
|
if (timeSpan > MAX_REDIS_REQUEST_LENGTH) {
|
|
error = new Error('redis getPreviousDocOps exceeded timeout')
|
|
return callback(error)
|
|
}
|
|
callback(null, ops)
|
|
}
|
|
)
|
|
})
|
|
})
|
|
},
|
|
|
|
DOC_OPS_TTL: 60 * minutes,
|
|
DOC_OPS_MAX_LENGTH: 100,
|
|
updateDocument(
|
|
projectId,
|
|
docId,
|
|
docLines,
|
|
newVersion,
|
|
appliedOps,
|
|
ranges,
|
|
updateMeta,
|
|
callback
|
|
) {
|
|
if (appliedOps == null) {
|
|
appliedOps = []
|
|
}
|
|
RedisManager.getDocVersion(docId, (error, currentVersion) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
if (currentVersion + appliedOps.length !== newVersion) {
|
|
error = new Error(`Version mismatch. '${docId}' is corrupted.`)
|
|
logger.error(
|
|
{
|
|
err: error,
|
|
docId,
|
|
currentVersion,
|
|
newVersion,
|
|
opsLength: appliedOps.length,
|
|
},
|
|
'version mismatch'
|
|
)
|
|
return callback(error)
|
|
}
|
|
|
|
const jsonOps = appliedOps.map(op => JSON.stringify(op))
|
|
for (const op of jsonOps) {
|
|
if (op.indexOf('\u0000') !== -1) {
|
|
error = new Error('null bytes found in jsonOps')
|
|
// this check was added to catch memory corruption in JSON.stringify
|
|
logger.error({ err: error, docId, jsonOps }, error.message)
|
|
return callback(error)
|
|
}
|
|
}
|
|
|
|
const newDocLines = JSON.stringify(docLines)
|
|
if (newDocLines.indexOf('\u0000') !== -1) {
|
|
error = new Error('null bytes found in doc lines')
|
|
// this check was added to catch memory corruption in JSON.stringify
|
|
logger.error({ err: error, docId, newDocLines }, error.message)
|
|
return callback(error)
|
|
}
|
|
// Do an optimised size check on the docLines using the serialised
|
|
// length as an upper bound
|
|
const sizeBound = newDocLines.length
|
|
if (docIsTooLarge(sizeBound, docLines, Settings.max_doc_length)) {
|
|
const err = new Error('blocking doc update: doc is too large')
|
|
const docSize = newDocLines.length
|
|
logger.error({ projectId, docId, err, docSize }, err.message)
|
|
return callback(err)
|
|
}
|
|
const newHash = RedisManager._computeHash(newDocLines)
|
|
|
|
const opVersions = appliedOps.map(op => op?.v)
|
|
logger.debug(
|
|
{
|
|
docId,
|
|
version: newVersion,
|
|
hash: newHash,
|
|
opVersions,
|
|
},
|
|
'updating doc in redis'
|
|
)
|
|
// record bytes sent to redis in update
|
|
metrics.summary('redis.docLines', newDocLines.length, {
|
|
status: 'update',
|
|
})
|
|
RedisManager._serializeRanges(ranges, (error, ranges) => {
|
|
if (error) {
|
|
logger.error({ err: error, docId }, error.message)
|
|
return callback(error)
|
|
}
|
|
if (ranges && ranges.indexOf('\u0000') !== -1) {
|
|
error = new Error('null bytes found in ranges')
|
|
// this check was added to catch memory corruption in JSON.stringify
|
|
logger.error({ err: error, docId, ranges }, error.message)
|
|
return callback(error)
|
|
}
|
|
|
|
// Make sure that this MULTI operation only operates on doc
|
|
// specific keys, i.e. keys that have the doc id in curly braces.
|
|
// The curly braces identify a hash key for Redis and ensures that
|
|
// the MULTI's operations are all done on the same node in a
|
|
// cluster environment.
|
|
const multi = rclient.multi()
|
|
multi.mset({
|
|
[keys.docLines({ doc_id: docId })]: newDocLines,
|
|
[keys.docVersion({ doc_id: docId })]: newVersion,
|
|
[keys.docHash({ doc_id: docId })]: newHash,
|
|
[keys.ranges({ doc_id: docId })]: ranges,
|
|
[keys.lastUpdatedAt({ doc_id: docId })]: Date.now(),
|
|
[keys.lastUpdatedBy({ doc_id: docId })]:
|
|
updateMeta && updateMeta.user_id,
|
|
})
|
|
multi.ltrim(
|
|
keys.docOps({ doc_id: docId }),
|
|
-RedisManager.DOC_OPS_MAX_LENGTH,
|
|
-1
|
|
) // index 3
|
|
// push the ops last so we can get the lengths at fixed index position 7
|
|
if (jsonOps.length > 0) {
|
|
multi.rpush(keys.docOps({ doc_id: docId }), ...jsonOps) // index 5
|
|
// expire must come after rpush since before it will be a no-op if the list is empty
|
|
multi.expire(keys.docOps({ doc_id: docId }), RedisManager.DOC_OPS_TTL) // index 6
|
|
}
|
|
// Set the unflushed timestamp to the current time if not set ("NX" flag).
|
|
multi.set(keys.unflushedTime({ doc_id: docId }), Date.now(), 'NX')
|
|
multi.exec((error, result) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
callback()
|
|
})
|
|
})
|
|
})
|
|
},
|
|
|
|
renameDoc(projectId, docId, userId, update, projectHistoryId, callback) {
|
|
RedisManager.getDoc(projectId, docId, (error, lines, version) => {
|
|
if (error) {
|
|
return callback(error)
|
|
}
|
|
if (lines != null && version != null) {
|
|
if (!update.newPathname) {
|
|
logger.warn(
|
|
{ projectId, docId, update },
|
|
'missing pathname in RedisManager.renameDoc'
|
|
)
|
|
metrics.inc('pathname', 1, {
|
|
path: 'RedisManager.renameDoc',
|
|
status: update.newPathname === '' ? 'zero-length' : 'undefined',
|
|
})
|
|
}
|
|
rclient.set(
|
|
keys.pathname({ doc_id: docId }),
|
|
update.newPathname,
|
|
callback
|
|
)
|
|
} else {
|
|
callback()
|
|
}
|
|
})
|
|
},
|
|
|
|
clearUnflushedTime(docId, callback) {
|
|
rclient.del(keys.unflushedTime({ doc_id: docId }), callback)
|
|
},
|
|
|
|
updateCommentState(docId, commentId, resolved, callback) {
|
|
if (resolved) {
|
|
rclient.sadd(
|
|
keys.resolvedCommentIds({ doc_id: docId }),
|
|
commentId,
|
|
callback
|
|
)
|
|
} else {
|
|
rclient.srem(
|
|
keys.resolvedCommentIds({ doc_id: docId }),
|
|
commentId,
|
|
callback
|
|
)
|
|
}
|
|
},
|
|
|
|
getDocIdsInProject(projectId, callback) {
|
|
rclient.smembers(keys.docsInProject({ project_id: projectId }), callback)
|
|
},
|
|
|
|
/**
|
|
* Get lastupdatedat timestamps for an array of docIds
|
|
*/
|
|
getDocTimestamps(docIds, callback) {
|
|
async.mapSeries(
|
|
docIds,
|
|
(docId, cb) => rclient.get(keys.lastUpdatedAt({ doc_id: docId }), cb),
|
|
callback
|
|
)
|
|
},
|
|
|
|
/**
|
|
* Store the project id in a sorted set ordered by time with a random offset
|
|
* to smooth out spikes
|
|
*/
|
|
queueFlushAndDeleteProject(projectId, callback) {
|
|
const SMOOTHING_OFFSET =
|
|
Settings.smoothingOffset > 0
|
|
? Math.round(Settings.smoothingOffset * Math.random())
|
|
: 0
|
|
rclient.zadd(
|
|
keys.flushAndDeleteQueue(),
|
|
Date.now() + SMOOTHING_OFFSET,
|
|
projectId,
|
|
callback
|
|
)
|
|
},
|
|
|
|
/**
|
|
* Find the oldest queued flush that is before the cutoff time
|
|
*/
|
|
getNextProjectToFlushAndDelete(cutoffTime, callback) {
|
|
rclient.zrangebyscore(
|
|
keys.flushAndDeleteQueue(),
|
|
0,
|
|
cutoffTime,
|
|
'WITHSCORES',
|
|
'LIMIT',
|
|
0,
|
|
1,
|
|
(err, reply) => {
|
|
if (err) {
|
|
return callback(err)
|
|
}
|
|
// return if no projects ready to be processed
|
|
if (!reply || reply.length === 0) {
|
|
return callback()
|
|
}
|
|
// pop the oldest entry (get and remove in a multi)
|
|
const multi = rclient.multi()
|
|
// Poor man's version of ZPOPMIN, which is only available in Redis 5.
|
|
multi.zrange(keys.flushAndDeleteQueue(), 0, 0, 'WITHSCORES')
|
|
multi.zremrangebyrank(keys.flushAndDeleteQueue(), 0, 0)
|
|
multi.zcard(keys.flushAndDeleteQueue()) // the total length of the queue (for metrics)
|
|
multi.exec((err, reply) => {
|
|
if (err) {
|
|
return callback(err)
|
|
}
|
|
if (!reply || reply.length === 0) {
|
|
return callback()
|
|
}
|
|
const [key, timestamp] = reply[0]
|
|
const queueLength = reply[2]
|
|
callback(null, key, timestamp, queueLength)
|
|
})
|
|
}
|
|
)
|
|
},
|
|
|
|
setHistoryRangesSupportFlag(docId, historyRangesSupport, callback) {
|
|
if (historyRangesSupport) {
|
|
rclient.sadd(keys.historyRangesSupport(), docId, callback)
|
|
} else {
|
|
rclient.srem(keys.historyRangesSupport(), docId, callback)
|
|
}
|
|
},
|
|
|
|
blockProject(projectId, callback) {
|
|
// Make sure that this MULTI operation only operates on project
|
|
// specific keys, i.e. keys that have the project id in curly braces.
|
|
// The curly braces identify a hash key for Redis and ensures that
|
|
// the MULTI's operations are all done on the same node in a
|
|
// cluster environment.
|
|
const multi = rclient.multi()
|
|
multi.setex(
|
|
keys.projectBlock({ project_id: projectId }),
|
|
PROJECT_BLOCK_TTL_SECS,
|
|
'1'
|
|
)
|
|
multi.scard(keys.docsInProject({ project_id: projectId }))
|
|
multi.exec((err, reply) => {
|
|
if (err) {
|
|
return callback(err)
|
|
}
|
|
const docsInProject = reply[1]
|
|
if (docsInProject > 0) {
|
|
// Too late to lock the project
|
|
rclient.del(keys.projectBlock({ project_id: projectId }), err => {
|
|
if (err) {
|
|
return callback(err)
|
|
}
|
|
callback(null, false)
|
|
})
|
|
} else {
|
|
callback(null, true)
|
|
}
|
|
})
|
|
},
|
|
|
|
unblockProject(projectId, callback) {
|
|
rclient.del(keys.projectBlock({ project_id: projectId }), (err, reply) => {
|
|
if (err) {
|
|
return callback(err)
|
|
}
|
|
const wasBlocked = reply === 1
|
|
callback(null, wasBlocked)
|
|
})
|
|
},
|
|
|
|
_serializeRanges(ranges, callback) {
|
|
let jsonRanges = JSON.stringify(ranges)
|
|
if (jsonRanges && jsonRanges.length > MAX_RANGES_SIZE) {
|
|
return callback(new Error('ranges are too large'))
|
|
}
|
|
if (jsonRanges === '{}') {
|
|
// Most doc will have empty ranges so don't fill redis with lots of '{}' keys
|
|
jsonRanges = null
|
|
}
|
|
callback(null, jsonRanges)
|
|
},
|
|
|
|
_deserializeRanges(ranges) {
|
|
if (ranges == null || ranges === '') {
|
|
return {}
|
|
} else {
|
|
return JSON.parse(ranges)
|
|
}
|
|
},
|
|
|
|
_computeHash(docLines) {
|
|
// use sha1 checksum of doclines to detect data corruption.
|
|
//
|
|
// note: must specify 'utf8' encoding explicitly, as the default is
|
|
// binary in node < v5
|
|
return crypto.createHash('sha1').update(docLines, 'utf8').digest('hex')
|
|
},
|
|
}
|
|
|
|
module.exports = RedisManager
|
|
module.exports.promises = promisifyAll(RedisManager, {
|
|
without: ['_deserializeRanges', '_computeHash'],
|
|
multiResult: {
|
|
getDoc: [
|
|
'lines',
|
|
'version',
|
|
'ranges',
|
|
'pathname',
|
|
'projectHistoryId',
|
|
'unflushedTime',
|
|
'lastUpdatedAt',
|
|
'lastUpdatedBy',
|
|
'historyRangesSupport',
|
|
'resolvedCommentIds',
|
|
],
|
|
getNextProjectToFlushAndDelete: [
|
|
'projectId',
|
|
'flushTimestamp',
|
|
'queueLength',
|
|
],
|
|
},
|
|
})
|