mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-07 20:31:06 -05:00
8bde496da4
* Send operations to project-history when resolving/unresolving comments * small fixes * added doc_id in web unit test * Revert "added doc_id in web unit test" This reverts commit f0b8251abfce17965d5e1b0e45d8784fcf1d9eed. * fix mocked dependency in test * wip: web unit tests * document updater, reopen test * document-updater tests * format fix * fix typo * fix callsArgWith * fix reopenThread calls in doc updater tests * fix typos * log error if chat api resolve failes * log error when reopening thread * sendStatus calls done() in tests * using OError instead of logging * removed timers * preserve legacy endpoints * update after merge * Remove timer check in HttpControllerTest * prettier * added "legacy" in log * remove metrics.timer * fix promisify issues * remove unused cb GitOrigin-RevId: 849538c86996973a065c727835e93028e5429344
292 lines
8.2 KiB
JavaScript
292 lines
8.2 KiB
JavaScript
// Metrics must be initialized before importing anything else
|
|
require('@overleaf/metrics/initialize')
|
|
|
|
const Metrics = require('@overleaf/metrics')
|
|
const express = require('express')
|
|
const Settings = require('@overleaf/settings')
|
|
const logger = require('@overleaf/logger')
|
|
logger.initialize('document-updater')
|
|
|
|
logger.logger.addSerializers(require('./app/js/LoggerSerializers'))
|
|
|
|
if (Settings.sentry != null && Settings.sentry.dsn != null) {
|
|
logger.initializeErrorReporting(Settings.sentry.dsn)
|
|
}
|
|
|
|
const RedisManager = require('./app/js/RedisManager')
|
|
const DispatchManager = require('./app/js/DispatchManager')
|
|
const DeleteQueueManager = require('./app/js/DeleteQueueManager')
|
|
const Errors = require('./app/js/Errors')
|
|
const HttpController = require('./app/js/HttpController')
|
|
const mongodb = require('./app/js/mongodb')
|
|
const async = require('async')
|
|
|
|
const bodyParser = require('body-parser')
|
|
|
|
Metrics.event_loop.monitor(logger, 100)
|
|
Metrics.open_sockets.monitor()
|
|
|
|
const app = express()
|
|
app.use(bodyParser.json({ limit: Settings.maxJsonRequestSize }))
|
|
Metrics.injectMetricsRoute(app)
|
|
|
|
DispatchManager.createAndStartDispatchers(Settings.dispatcherCount)
|
|
|
|
app.get('/status', (req, res) => {
|
|
if (Settings.shuttingDown) {
|
|
return res.sendStatus(503) // Service unavailable
|
|
} else {
|
|
return res.send('document updater is alive')
|
|
}
|
|
})
|
|
|
|
const pubsubClient = require('@overleaf/redis-wrapper').createClient(
|
|
Settings.redis.pubsub
|
|
)
|
|
app.get('/health_check/redis', (req, res, next) => {
|
|
pubsubClient.healthCheck(error => {
|
|
if (error) {
|
|
logger.err({ err: error }, 'failed redis health check')
|
|
return res.sendStatus(500)
|
|
} else {
|
|
return res.sendStatus(200)
|
|
}
|
|
})
|
|
})
|
|
|
|
const docUpdaterRedisClient = require('@overleaf/redis-wrapper').createClient(
|
|
Settings.redis.documentupdater
|
|
)
|
|
app.get('/health_check/redis_cluster', (req, res, next) => {
|
|
docUpdaterRedisClient.healthCheck(error => {
|
|
if (error) {
|
|
logger.err({ err: error }, 'failed redis cluster health check')
|
|
return res.sendStatus(500)
|
|
} else {
|
|
return res.sendStatus(200)
|
|
}
|
|
})
|
|
})
|
|
|
|
app.get('/health_check', (req, res, next) => {
|
|
async.series(
|
|
[
|
|
cb => {
|
|
pubsubClient.healthCheck(error => {
|
|
if (error) {
|
|
logger.err({ err: error }, 'failed redis health check')
|
|
}
|
|
cb(error)
|
|
})
|
|
},
|
|
cb => {
|
|
docUpdaterRedisClient.healthCheck(error => {
|
|
if (error) {
|
|
logger.err({ err: error }, 'failed redis cluster health check')
|
|
}
|
|
cb(error)
|
|
})
|
|
},
|
|
cb => {
|
|
mongodb.healthCheck(error => {
|
|
if (error) {
|
|
logger.err({ err: error }, 'failed mongo health check')
|
|
}
|
|
cb(error)
|
|
})
|
|
},
|
|
],
|
|
error => {
|
|
if (error) {
|
|
return res.sendStatus(500)
|
|
} else {
|
|
return res.sendStatus(200)
|
|
}
|
|
}
|
|
)
|
|
})
|
|
|
|
// record http metrics for the routes below this point
|
|
app.use(Metrics.http.monitor(logger))
|
|
|
|
app.param('project_id', (req, res, next, projectId) => {
|
|
if (projectId != null && projectId.match(/^[0-9a-f]{24}$/)) {
|
|
return next()
|
|
} else {
|
|
return next(new Error('invalid project id'))
|
|
}
|
|
})
|
|
|
|
app.param('doc_id', (req, res, next, docId) => {
|
|
if (docId != null && docId.match(/^[0-9a-f]{24}$/)) {
|
|
return next()
|
|
} else {
|
|
return next(new Error('invalid doc id'))
|
|
}
|
|
})
|
|
|
|
// Record requests that come in after we've started shutting down - for investigation.
|
|
app.use((req, res, next) => {
|
|
if (Settings.shuttingDown) {
|
|
logger.warn(
|
|
{ req, timeSinceShutdown: Date.now() - Settings.shutDownTime },
|
|
'request received after shutting down'
|
|
)
|
|
// We don't want keep-alive connections to be kept open when the server is shutting down.
|
|
res.set('Connection', 'close')
|
|
}
|
|
next()
|
|
})
|
|
|
|
app.get('/project/:project_id/doc/:doc_id', HttpController.getDoc)
|
|
app.get('/project/:project_id/doc/:doc_id/peek', HttpController.peekDoc)
|
|
// temporarily keep the GET method for backwards compatibility
|
|
app.get('/project/:project_id/doc', HttpController.getProjectDocsAndFlushIfOld)
|
|
// will migrate to the POST method of get_and_flush_if_old instead
|
|
app.post(
|
|
'/project/:project_id/get_and_flush_if_old',
|
|
HttpController.getProjectDocsAndFlushIfOld
|
|
)
|
|
app.post('/project/:project_id/clearState', HttpController.clearProjectState)
|
|
app.post('/project/:project_id/doc/:doc_id', HttpController.setDoc)
|
|
app.post(
|
|
'/project/:project_id/doc/:doc_id/flush',
|
|
HttpController.flushDocIfLoaded
|
|
)
|
|
app.delete('/project/:project_id/doc/:doc_id', HttpController.deleteDoc)
|
|
app.delete('/project/:project_id', HttpController.deleteProject)
|
|
app.delete('/project', HttpController.deleteMultipleProjects)
|
|
app.post('/project/:project_id', HttpController.updateProject)
|
|
app.post(
|
|
'/project/:project_id/history/resync',
|
|
longerTimeout,
|
|
HttpController.resyncProjectHistory
|
|
)
|
|
app.post('/project/:project_id/flush', HttpController.flushProject)
|
|
app.post(
|
|
'/project/:project_id/doc/:doc_id/change/:change_id/accept',
|
|
HttpController.acceptChanges
|
|
)
|
|
app.post(
|
|
'/project/:project_id/doc/:doc_id/change/accept',
|
|
HttpController.acceptChanges
|
|
)
|
|
app.post(
|
|
'/project/:project_id/doc/:doc_id/comment/:comment_id/resolve',
|
|
HttpController.resolveComment
|
|
)
|
|
app.post(
|
|
'/project/:project_id/doc/:doc_id/comment/:comment_id/reopen',
|
|
HttpController.reopenComment
|
|
)
|
|
app.delete(
|
|
'/project/:project_id/doc/:doc_id/comment/:comment_id',
|
|
HttpController.deleteComment
|
|
)
|
|
|
|
app.get('/flush_all_projects', HttpController.flushAllProjects)
|
|
app.get('/flush_queued_projects', HttpController.flushQueuedProjects)
|
|
|
|
app.get('/total', (req, res, next) => {
|
|
const timer = new Metrics.Timer('http.allDocList')
|
|
RedisManager.getCountOfDocsInMemory((err, count) => {
|
|
if (err) {
|
|
return next(err)
|
|
}
|
|
timer.done()
|
|
res.send({ total: count })
|
|
})
|
|
})
|
|
|
|
app.use((error, req, res, next) => {
|
|
if (error instanceof Errors.NotFoundError) {
|
|
return res.sendStatus(404)
|
|
} else if (error instanceof Errors.OpRangeNotAvailableError) {
|
|
return res.sendStatus(422) // Unprocessable Entity
|
|
} else if (error instanceof Errors.FileTooLargeError) {
|
|
return res.sendStatus(413)
|
|
} else if (error.statusCode === 413) {
|
|
return res.status(413).send('request entity too large')
|
|
} else {
|
|
logger.error({ err: error, req }, 'request errored')
|
|
return res.status(500).send('Oops, something went wrong')
|
|
}
|
|
})
|
|
|
|
const shutdownCleanly = signal => () => {
|
|
logger.info({ signal }, 'received interrupt, cleaning up')
|
|
if (Settings.shuttingDown) {
|
|
logger.warn({ signal }, 'already shutting down, ignoring interrupt')
|
|
return
|
|
}
|
|
Settings.shuttingDown = true
|
|
// record the time we started shutting down
|
|
Settings.shutDownTime = Date.now()
|
|
setTimeout(() => {
|
|
logger.info({ signal }, 'shutting down')
|
|
process.exit()
|
|
}, Settings.gracefulShutdownDelayInMs)
|
|
}
|
|
|
|
const watchForEvent = eventName => {
|
|
docUpdaterRedisClient.on(eventName, e => {
|
|
console.log(`redis event: ${eventName} ${e}`) // eslint-disable-line no-console
|
|
})
|
|
}
|
|
|
|
const events = ['connect', 'ready', 'error', 'close', 'reconnecting', 'end']
|
|
for (const eventName of events) {
|
|
watchForEvent(eventName)
|
|
}
|
|
|
|
const port =
|
|
Settings.internal.documentupdater.port ||
|
|
(Settings.api &&
|
|
Settings.api.documentupdater &&
|
|
Settings.api.documentupdater.port) ||
|
|
3003
|
|
const host = Settings.internal.documentupdater.host || 'localhost'
|
|
|
|
if (!module.parent) {
|
|
// Called directly
|
|
mongodb.mongoClient
|
|
.connect()
|
|
.then(() => {
|
|
app.listen(port, host, function (err) {
|
|
if (err) {
|
|
logger.fatal({ err }, `Cannot bind to ${host}:${port}. Exiting.`)
|
|
process.exit(1)
|
|
}
|
|
logger.info(
|
|
`Document-updater starting up, listening on ${host}:${port}`
|
|
)
|
|
if (Settings.continuousBackgroundFlush) {
|
|
logger.info('Starting continuous background flush')
|
|
DeleteQueueManager.startBackgroundFlush()
|
|
}
|
|
})
|
|
})
|
|
.catch(err => {
|
|
logger.fatal({ err }, 'Cannot connect to mongo. Exiting.')
|
|
process.exit(1)
|
|
})
|
|
}
|
|
|
|
module.exports = app
|
|
|
|
for (const signal of [
|
|
'SIGINT',
|
|
'SIGHUP',
|
|
'SIGQUIT',
|
|
'SIGUSR1',
|
|
'SIGUSR2',
|
|
'SIGTERM',
|
|
'SIGABRT',
|
|
]) {
|
|
process.on(signal, shutdownCleanly(signal))
|
|
}
|
|
|
|
function longerTimeout(req, res, next) {
|
|
res.setTimeout(6 * 60 * 1000)
|
|
next()
|
|
}
|