mirror of
https://github.com/overleaf/overleaf.git
synced 2024-12-27 05:03:39 +00:00
182 lines
6.2 KiB
CoffeeScript
182 lines
6.2 KiB
CoffeeScript
Metrics = require("metrics-sharelatex")
|
|
Settings = require "settings-sharelatex"
|
|
Metrics.initialize(Settings.appName or "real-time")
|
|
async = require("async")
|
|
_ = require "underscore"
|
|
|
|
logger = require "logger-sharelatex"
|
|
logger.initialize("real-time")
|
|
Metrics.event_loop.monitor(logger)
|
|
|
|
express = require("express")
|
|
session = require("express-session")
|
|
redis = require("redis-sharelatex")
|
|
if Settings.sentry?.dsn?
|
|
logger.initializeErrorReporting(Settings.sentry.dsn)
|
|
|
|
sessionRedisClient = redis.createClient(Settings.redis.websessions)
|
|
|
|
RedisStore = require('connect-redis')(session)
|
|
SessionSockets = require('./app/js/SessionSockets')
|
|
CookieParser = require("cookie-parser")
|
|
|
|
DrainManager = require("./app/js/DrainManager")
|
|
HealthCheckManager = require("./app/js/HealthCheckManager")
|
|
|
|
# work around frame handler bug in socket.io v0.9.16
|
|
require("./socket.io.patch.js")
|
|
# Set up socket.io server
|
|
app = express()
|
|
|
|
server = require('http').createServer(app)
|
|
io = require('socket.io').listen(server)
|
|
|
|
# Bind to sessions
|
|
sessionStore = new RedisStore(client: sessionRedisClient)
|
|
cookieParser = CookieParser(Settings.security.sessionSecret)
|
|
|
|
sessionSockets = new SessionSockets(io, sessionStore, cookieParser, Settings.cookieName)
|
|
|
|
Metrics.injectMetricsRoute(app)
|
|
app.use(Metrics.http.monitor(logger))
|
|
|
|
io.configure ->
|
|
io.enable('browser client minification')
|
|
io.enable('browser client etag')
|
|
|
|
# Fix for Safari 5 error of "Error during WebSocket handshake: location mismatch"
|
|
# See http://answers.dotcloud.com/question/578/problem-with-websocket-over-ssl-in-safari-with
|
|
io.set('match origin protocol', true)
|
|
|
|
# gzip uses a Node 0.8.x method of calling the gzip program which
|
|
# doesn't work with 0.6.x
|
|
#io.enable('browser client gzip')
|
|
io.set('transports', ['websocket', 'flashsocket', 'htmlfile', 'xhr-polling', 'jsonp-polling'])
|
|
io.set('log level', 1)
|
|
|
|
app.get "/", (req, res, next) ->
|
|
res.send "real-time-sharelatex is alive"
|
|
|
|
app.get "/status", (req, res, next) ->
|
|
if Settings.shutDownInProgress
|
|
res.send 503 # Service unavailable
|
|
else
|
|
res.send "real-time-sharelatex is alive"
|
|
|
|
app.get "/debug/events", (req, res, next) ->
|
|
Settings.debugEvents = parseInt(req.query?.count,10) || 20
|
|
logger.log {count: Settings.debugEvents}, "starting debug mode"
|
|
res.send "debug mode will log next #{Settings.debugEvents} events"
|
|
|
|
rclient = require("redis-sharelatex").createClient(Settings.redis.realtime)
|
|
|
|
healthCheck = (req, res, next)->
|
|
rclient.healthCheck (error) ->
|
|
if error?
|
|
logger.err {err: error}, "failed redis health check"
|
|
res.sendStatus 500
|
|
else if HealthCheckManager.isFailing()
|
|
status = HealthCheckManager.status()
|
|
logger.err {pubSubErrors: status}, "failed pubsub health check"
|
|
res.sendStatus 500
|
|
else
|
|
res.sendStatus 200
|
|
|
|
app.get "/health_check", healthCheck
|
|
|
|
app.get "/health_check/redis", healthCheck
|
|
|
|
|
|
|
|
Router = require "./app/js/Router"
|
|
Router.configure(app, io, sessionSockets)
|
|
|
|
WebsocketLoadBalancer = require "./app/js/WebsocketLoadBalancer"
|
|
WebsocketLoadBalancer.listenForEditorEvents(io)
|
|
|
|
DocumentUpdaterController = require "./app/js/DocumentUpdaterController"
|
|
DocumentUpdaterController.listenForUpdatesFromDocumentUpdater(io)
|
|
|
|
port = Settings.internal.realTime.port
|
|
host = Settings.internal.realTime.host
|
|
|
|
server.listen port, host, (error) ->
|
|
throw error if error?
|
|
logger.info "realtime starting up, listening on #{host}:#{port}"
|
|
|
|
# Stop huge stack traces in logs from all the socket.io parsing steps.
|
|
Error.stackTraceLimit = 10
|
|
|
|
|
|
shutdownCleanly = (signal) ->
|
|
connectedClients = io.sockets.clients()?.length
|
|
if connectedClients == 0
|
|
logger.warn("no clients connected, exiting")
|
|
process.exit()
|
|
else
|
|
logger.warn {connectedClients}, "clients still connected, not shutting down yet"
|
|
setTimeout () ->
|
|
shutdownCleanly(signal)
|
|
, 30 * 1000
|
|
|
|
drainAndShutdown = (signal) ->
|
|
if Settings.shutDownInProgress
|
|
logger.warn signal: signal, "shutdown already in progress, ignoring signal"
|
|
return
|
|
else
|
|
Settings.shutDownInProgress = true
|
|
statusCheckInterval = Settings.statusCheckInterval
|
|
if statusCheckInterval
|
|
logger.warn signal: signal, "received interrupt, delay drain by #{statusCheckInterval}ms"
|
|
setTimeout () ->
|
|
logger.warn signal: signal, "received interrupt, starting drain over #{shutdownDrainTimeWindow} mins"
|
|
DrainManager.startDrainTimeWindow(io, shutdownDrainTimeWindow)
|
|
shutdownCleanly(signal)
|
|
, statusCheckInterval
|
|
|
|
|
|
Settings.shutDownInProgress = false
|
|
if Settings.shutdownDrainTimeWindow?
|
|
shutdownDrainTimeWindow = parseInt(Settings.shutdownDrainTimeWindow, 10)
|
|
logger.log shutdownDrainTimeWindow: shutdownDrainTimeWindow,"shutdownDrainTimeWindow enabled"
|
|
for signal in ['SIGINT', 'SIGHUP', 'SIGQUIT', 'SIGUSR1', 'SIGUSR2', 'SIGTERM', 'SIGABRT']
|
|
process.on signal, drainAndShutdown # signal is passed as argument to event handler
|
|
|
|
# global exception handler
|
|
if Settings.errors?.catchUncaughtErrors
|
|
process.removeAllListeners('uncaughtException')
|
|
process.on 'uncaughtException', (error) ->
|
|
if ['EPIPE', 'ECONNRESET'].includes(error.code)
|
|
Metrics.inc('disconnected_write', 1, {status: error.code})
|
|
return logger.warn err: error, 'attempted to write to disconnected client'
|
|
logger.error err: error, 'uncaught exception'
|
|
if Settings.errors?.shutdownOnUncaughtError
|
|
drainAndShutdown('SIGABRT')
|
|
|
|
if Settings.continualPubsubTraffic
|
|
console.log "continualPubsubTraffic enabled"
|
|
|
|
pubsubClient = redis.createClient(Settings.redis.pubsub)
|
|
clusterClient = redis.createClient(Settings.redis.websessions)
|
|
|
|
publishJob = (channel, callback)->
|
|
checker = new HealthCheckManager(channel)
|
|
logger.debug {channel:channel}, "sending pub to keep connection alive"
|
|
json = JSON.stringify({health_check:true, key: checker.id, date: new Date().toString()})
|
|
Metrics.summary "redis.publish.#{channel}", json.length
|
|
pubsubClient.publish channel, json, (err)->
|
|
if err?
|
|
logger.err {err, channel}, "error publishing pubsub traffic to redis"
|
|
blob = JSON.stringify({keep: "alive"})
|
|
Metrics.summary "redis.publish.cluster-continual-traffic", blob.length
|
|
clusterClient.publish "cluster-continual-traffic", blob, callback
|
|
|
|
|
|
runPubSubTraffic = ->
|
|
async.map ["applied-ops", "editor-events"], publishJob, (err)->
|
|
setTimeout(runPubSubTraffic, 1000 * 20)
|
|
|
|
runPubSubTraffic()
|
|
|
|
|
|
|