mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-21 20:47:08 -05:00
decaffeinate: convert individual files to js
This commit is contained in:
parent
66ce5847a3
commit
d23250a4bb
2 changed files with 311 additions and 239 deletions
|
@ -1,244 +1,298 @@
|
|||
Metrics = require "metrics-sharelatex"
|
||||
Metrics.initialize("clsi")
|
||||
/*
|
||||
* decaffeinate suggestions:
|
||||
* DS102: Remove unnecessary code created because of implicit returns
|
||||
* DS103: Rewrite code to no longer use __guard__
|
||||
* DS207: Consider shorter variations of null checks
|
||||
* Full docs: https://github.com/decaffeinate/decaffeinate/blob/master/docs/suggestions.md
|
||||
*/
|
||||
let tenMinutes;
|
||||
const Metrics = require("metrics-sharelatex");
|
||||
Metrics.initialize("clsi");
|
||||
|
||||
CompileController = require "./app/js/CompileController"
|
||||
Settings = require "settings-sharelatex"
|
||||
logger = require "logger-sharelatex"
|
||||
logger.initialize("clsi")
|
||||
if Settings.sentry?.dsn?
|
||||
logger.initializeErrorReporting(Settings.sentry.dsn)
|
||||
const CompileController = require("./app/js/CompileController");
|
||||
const Settings = require("settings-sharelatex");
|
||||
const logger = require("logger-sharelatex");
|
||||
logger.initialize("clsi");
|
||||
if ((Settings.sentry != null ? Settings.sentry.dsn : undefined) != null) {
|
||||
logger.initializeErrorReporting(Settings.sentry.dsn);
|
||||
}
|
||||
|
||||
smokeTest = require "smoke-test-sharelatex"
|
||||
ContentTypeMapper = require "./app/js/ContentTypeMapper"
|
||||
Errors = require './app/js/Errors'
|
||||
const smokeTest = require("smoke-test-sharelatex");
|
||||
const ContentTypeMapper = require("./app/js/ContentTypeMapper");
|
||||
const Errors = require('./app/js/Errors');
|
||||
|
||||
Path = require "path"
|
||||
fs = require "fs"
|
||||
const Path = require("path");
|
||||
const fs = require("fs");
|
||||
|
||||
|
||||
Metrics.open_sockets.monitor(logger)
|
||||
Metrics.memory.monitor(logger)
|
||||
Metrics.open_sockets.monitor(logger);
|
||||
Metrics.memory.monitor(logger);
|
||||
|
||||
ProjectPersistenceManager = require "./app/js/ProjectPersistenceManager"
|
||||
OutputCacheManager = require "./app/js/OutputCacheManager"
|
||||
const ProjectPersistenceManager = require("./app/js/ProjectPersistenceManager");
|
||||
const OutputCacheManager = require("./app/js/OutputCacheManager");
|
||||
|
||||
require("./app/js/db").sync()
|
||||
require("./app/js/db").sync();
|
||||
|
||||
express = require "express"
|
||||
bodyParser = require "body-parser"
|
||||
app = express()
|
||||
const express = require("express");
|
||||
const bodyParser = require("body-parser");
|
||||
const app = express();
|
||||
|
||||
Metrics.injectMetricsRoute(app)
|
||||
app.use Metrics.http.monitor(logger)
|
||||
Metrics.injectMetricsRoute(app);
|
||||
app.use(Metrics.http.monitor(logger));
|
||||
|
||||
# Compile requests can take longer than the default two
|
||||
# minutes (including file download time), so bump up the
|
||||
# timeout a bit.
|
||||
TIMEOUT = 10 * 60 * 1000
|
||||
app.use (req, res, next) ->
|
||||
req.setTimeout TIMEOUT
|
||||
res.setTimeout TIMEOUT
|
||||
res.removeHeader("X-Powered-By")
|
||||
next()
|
||||
// Compile requests can take longer than the default two
|
||||
// minutes (including file download time), so bump up the
|
||||
// timeout a bit.
|
||||
const TIMEOUT = 10 * 60 * 1000;
|
||||
app.use(function(req, res, next) {
|
||||
req.setTimeout(TIMEOUT);
|
||||
res.setTimeout(TIMEOUT);
|
||||
res.removeHeader("X-Powered-By");
|
||||
return next();
|
||||
});
|
||||
|
||||
app.param 'project_id', (req, res, next, project_id) ->
|
||||
if project_id?.match /^[a-zA-Z0-9_-]+$/
|
||||
next()
|
||||
else
|
||||
next new Error("invalid project id")
|
||||
app.param('project_id', function(req, res, next, project_id) {
|
||||
if ((project_id != null ? project_id.match(/^[a-zA-Z0-9_-]+$/) : undefined)) {
|
||||
return next();
|
||||
} else {
|
||||
return next(new Error("invalid project id"));
|
||||
}
|
||||
});
|
||||
|
||||
app.param 'user_id', (req, res, next, user_id) ->
|
||||
if user_id?.match /^[0-9a-f]{24}$/
|
||||
next()
|
||||
else
|
||||
next new Error("invalid user id")
|
||||
app.param('user_id', function(req, res, next, user_id) {
|
||||
if ((user_id != null ? user_id.match(/^[0-9a-f]{24}$/) : undefined)) {
|
||||
return next();
|
||||
} else {
|
||||
return next(new Error("invalid user id"));
|
||||
}
|
||||
});
|
||||
|
||||
app.param 'build_id', (req, res, next, build_id) ->
|
||||
if build_id?.match OutputCacheManager.BUILD_REGEX
|
||||
next()
|
||||
else
|
||||
next new Error("invalid build id #{build_id}")
|
||||
app.param('build_id', function(req, res, next, build_id) {
|
||||
if ((build_id != null ? build_id.match(OutputCacheManager.BUILD_REGEX) : undefined)) {
|
||||
return next();
|
||||
} else {
|
||||
return next(new Error(`invalid build id ${build_id}`));
|
||||
}
|
||||
});
|
||||
|
||||
|
||||
app.post "/project/:project_id/compile", bodyParser.json(limit: Settings.compileSizeLimit), CompileController.compile
|
||||
app.post "/project/:project_id/compile/stop", CompileController.stopCompile
|
||||
app.delete "/project/:project_id", CompileController.clearCache
|
||||
app.post("/project/:project_id/compile", bodyParser.json({limit: Settings.compileSizeLimit}), CompileController.compile);
|
||||
app.post("/project/:project_id/compile/stop", CompileController.stopCompile);
|
||||
app.delete("/project/:project_id", CompileController.clearCache);
|
||||
|
||||
app.get "/project/:project_id/sync/code", CompileController.syncFromCode
|
||||
app.get "/project/:project_id/sync/pdf", CompileController.syncFromPdf
|
||||
app.get "/project/:project_id/wordcount", CompileController.wordcount
|
||||
app.get "/project/:project_id/status", CompileController.status
|
||||
app.get("/project/:project_id/sync/code", CompileController.syncFromCode);
|
||||
app.get("/project/:project_id/sync/pdf", CompileController.syncFromPdf);
|
||||
app.get("/project/:project_id/wordcount", CompileController.wordcount);
|
||||
app.get("/project/:project_id/status", CompileController.status);
|
||||
|
||||
# Per-user containers
|
||||
app.post "/project/:project_id/user/:user_id/compile", bodyParser.json(limit: Settings.compileSizeLimit), CompileController.compile
|
||||
app.post "/project/:project_id/user/:user_id/compile/stop", CompileController.stopCompile
|
||||
app.delete "/project/:project_id/user/:user_id", CompileController.clearCache
|
||||
// Per-user containers
|
||||
app.post("/project/:project_id/user/:user_id/compile", bodyParser.json({limit: Settings.compileSizeLimit}), CompileController.compile);
|
||||
app.post("/project/:project_id/user/:user_id/compile/stop", CompileController.stopCompile);
|
||||
app.delete("/project/:project_id/user/:user_id", CompileController.clearCache);
|
||||
|
||||
app.get "/project/:project_id/user/:user_id/sync/code", CompileController.syncFromCode
|
||||
app.get "/project/:project_id/user/:user_id/sync/pdf", CompileController.syncFromPdf
|
||||
app.get "/project/:project_id/user/:user_id/wordcount", CompileController.wordcount
|
||||
app.get("/project/:project_id/user/:user_id/sync/code", CompileController.syncFromCode);
|
||||
app.get("/project/:project_id/user/:user_id/sync/pdf", CompileController.syncFromPdf);
|
||||
app.get("/project/:project_id/user/:user_id/wordcount", CompileController.wordcount);
|
||||
|
||||
ForbidSymlinks = require "./app/js/StaticServerForbidSymlinks"
|
||||
const ForbidSymlinks = require("./app/js/StaticServerForbidSymlinks");
|
||||
|
||||
# create a static server which does not allow access to any symlinks
|
||||
# avoids possible mismatch of root directory between middleware check
|
||||
# and serving the files
|
||||
staticServer = ForbidSymlinks express.static, Settings.path.compilesDir, setHeaders: (res, path, stat) ->
|
||||
if Path.basename(path) == "output.pdf"
|
||||
# Calculate an etag in the same way as nginx
|
||||
# https://github.com/tj/send/issues/65
|
||||
etag = (path, stat) ->
|
||||
'"' + Math.ceil(+stat.mtime / 1000).toString(16) +
|
||||
// create a static server which does not allow access to any symlinks
|
||||
// avoids possible mismatch of root directory between middleware check
|
||||
// and serving the files
|
||||
const staticServer = ForbidSymlinks(express.static, Settings.path.compilesDir, { setHeaders(res, path, stat) {
|
||||
if (Path.basename(path) === "output.pdf") {
|
||||
// Calculate an etag in the same way as nginx
|
||||
// https://github.com/tj/send/issues/65
|
||||
const etag = (path, stat) =>
|
||||
`"${Math.ceil(+stat.mtime / 1000).toString(16)}` +
|
||||
'-' + Number(stat.size).toString(16) + '"'
|
||||
res.set("Etag", etag(path, stat))
|
||||
res.set("Content-Type", ContentTypeMapper.map(path))
|
||||
;
|
||||
res.set("Etag", etag(path, stat));
|
||||
}
|
||||
return res.set("Content-Type", ContentTypeMapper.map(path));
|
||||
}
|
||||
}
|
||||
);
|
||||
|
||||
app.get "/project/:project_id/user/:user_id/build/:build_id/output/*", (req, res, next) ->
|
||||
# for specific build get the path from the OutputCacheManager (e.g. .clsi/buildId)
|
||||
req.url = "/#{req.params.project_id}-#{req.params.user_id}/" + OutputCacheManager.path(req.params.build_id, "/#{req.params[0]}")
|
||||
staticServer(req, res, next)
|
||||
app.get("/project/:project_id/user/:user_id/build/:build_id/output/*", function(req, res, next) {
|
||||
// for specific build get the path from the OutputCacheManager (e.g. .clsi/buildId)
|
||||
req.url = `/${req.params.project_id}-${req.params.user_id}/` + OutputCacheManager.path(req.params.build_id, `/${req.params[0]}`);
|
||||
return staticServer(req, res, next);
|
||||
});
|
||||
|
||||
app.get "/project/:project_id/build/:build_id/output/*", (req, res, next) ->
|
||||
# for specific build get the path from the OutputCacheManager (e.g. .clsi/buildId)
|
||||
req.url = "/#{req.params.project_id}/" + OutputCacheManager.path(req.params.build_id, "/#{req.params[0]}")
|
||||
staticServer(req, res, next)
|
||||
app.get("/project/:project_id/build/:build_id/output/*", function(req, res, next) {
|
||||
// for specific build get the path from the OutputCacheManager (e.g. .clsi/buildId)
|
||||
req.url = `/${req.params.project_id}/` + OutputCacheManager.path(req.params.build_id, `/${req.params[0]}`);
|
||||
return staticServer(req, res, next);
|
||||
});
|
||||
|
||||
app.get "/project/:project_id/user/:user_id/output/*", (req, res, next) ->
|
||||
# for specific user get the path to the top level file
|
||||
req.url = "/#{req.params.project_id}-#{req.params.user_id}/#{req.params[0]}"
|
||||
staticServer(req, res, next)
|
||||
app.get("/project/:project_id/user/:user_id/output/*", function(req, res, next) {
|
||||
// for specific user get the path to the top level file
|
||||
req.url = `/${req.params.project_id}-${req.params.user_id}/${req.params[0]}`;
|
||||
return staticServer(req, res, next);
|
||||
});
|
||||
|
||||
app.get "/project/:project_id/output/*", (req, res, next) ->
|
||||
if req.query?.build? && req.query.build.match(OutputCacheManager.BUILD_REGEX)
|
||||
# for specific build get the path from the OutputCacheManager (e.g. .clsi/buildId)
|
||||
req.url = "/#{req.params.project_id}/" + OutputCacheManager.path(req.query.build, "/#{req.params[0]}")
|
||||
else
|
||||
req.url = "/#{req.params.project_id}/#{req.params[0]}"
|
||||
staticServer(req, res, next)
|
||||
app.get("/project/:project_id/output/*", function(req, res, next) {
|
||||
if (((req.query != null ? req.query.build : undefined) != null) && req.query.build.match(OutputCacheManager.BUILD_REGEX)) {
|
||||
// for specific build get the path from the OutputCacheManager (e.g. .clsi/buildId)
|
||||
req.url = `/${req.params.project_id}/` + OutputCacheManager.path(req.query.build, `/${req.params[0]}`);
|
||||
} else {
|
||||
req.url = `/${req.params.project_id}/${req.params[0]}`;
|
||||
}
|
||||
return staticServer(req, res, next);
|
||||
});
|
||||
|
||||
app.get "/oops", (req, res, next) ->
|
||||
logger.error {err: "hello"}, "test error"
|
||||
res.send "error\n"
|
||||
app.get("/oops", function(req, res, next) {
|
||||
logger.error({err: "hello"}, "test error");
|
||||
return res.send("error\n");
|
||||
});
|
||||
|
||||
|
||||
app.get "/status", (req, res, next) ->
|
||||
res.send "CLSI is alive\n"
|
||||
app.get("/status", (req, res, next) => res.send("CLSI is alive\n"));
|
||||
|
||||
resCacher =
|
||||
contentType:(@setContentType)->
|
||||
send:(@code, @body)->
|
||||
const resCacher = {
|
||||
contentType(setContentType){
|
||||
this.setContentType = setContentType;
|
||||
},
|
||||
send(code, body){
|
||||
this.code = code;
|
||||
this.body = body;
|
||||
},
|
||||
|
||||
#default the server to be down
|
||||
code:500
|
||||
body:{}
|
||||
//default the server to be down
|
||||
code:500,
|
||||
body:{},
|
||||
setContentType:"application/json"
|
||||
};
|
||||
|
||||
if Settings.smokeTest
|
||||
do runSmokeTest = ->
|
||||
logger.log("running smoke tests")
|
||||
smokeTest.run(require.resolve(__dirname + "/test/smoke/js/SmokeTests.js"))({}, resCacher)
|
||||
setTimeout(runSmokeTest, 30 * 1000)
|
||||
if (Settings.smokeTest) {
|
||||
let runSmokeTest;
|
||||
(runSmokeTest = function() {
|
||||
logger.log("running smoke tests");
|
||||
smokeTest.run(require.resolve(__dirname + "/test/smoke/js/SmokeTests.js"))({}, resCacher);
|
||||
return setTimeout(runSmokeTest, 30 * 1000);
|
||||
})();
|
||||
}
|
||||
|
||||
app.get "/health_check", (req, res)->
|
||||
res.contentType(resCacher?.setContentType)
|
||||
res.status(resCacher?.code).send(resCacher?.body)
|
||||
app.get("/health_check", function(req, res){
|
||||
res.contentType(resCacher != null ? resCacher.setContentType : undefined);
|
||||
return res.status(resCacher != null ? resCacher.code : undefined).send(resCacher != null ? resCacher.body : undefined);
|
||||
});
|
||||
|
||||
app.get "/smoke_test_force", (req, res)->
|
||||
smokeTest.run(require.resolve(__dirname + "/test/smoke/js/SmokeTests.js"))(req, res)
|
||||
app.get("/smoke_test_force", (req, res)=> smokeTest.run(require.resolve(__dirname + "/test/smoke/js/SmokeTests.js"))(req, res));
|
||||
|
||||
profiler = require "v8-profiler-node8"
|
||||
app.get "/profile", (req, res) ->
|
||||
time = parseInt(req.query.time || "1000")
|
||||
profiler.startProfiling("test")
|
||||
setTimeout () ->
|
||||
profile = profiler.stopProfiling("test")
|
||||
res.json(profile)
|
||||
, time
|
||||
const profiler = require("v8-profiler-node8");
|
||||
app.get("/profile", function(req, res) {
|
||||
const time = parseInt(req.query.time || "1000");
|
||||
profiler.startProfiling("test");
|
||||
return setTimeout(function() {
|
||||
const profile = profiler.stopProfiling("test");
|
||||
return res.json(profile);
|
||||
}
|
||||
, time);
|
||||
});
|
||||
|
||||
app.get "/heapdump", (req, res)->
|
||||
require('heapdump').writeSnapshot '/tmp/' + Date.now() + '.clsi.heapsnapshot', (err, filename)->
|
||||
res.send filename
|
||||
app.get("/heapdump", (req, res)=>
|
||||
require('heapdump').writeSnapshot(`/tmp/${Date.now()}.clsi.heapsnapshot`, (err, filename)=> res.send(filename))
|
||||
);
|
||||
|
||||
app.use (error, req, res, next) ->
|
||||
if error instanceof Errors.NotFoundError
|
||||
logger.warn {err: error, url: req.url}, "not found error"
|
||||
return res.sendStatus(404)
|
||||
else
|
||||
logger.error {err: error, url: req.url}, "server error"
|
||||
res.sendStatus(error?.statusCode || 500)
|
||||
app.use(function(error, req, res, next) {
|
||||
if (error instanceof Errors.NotFoundError) {
|
||||
logger.warn({err: error, url: req.url}, "not found error");
|
||||
return res.sendStatus(404);
|
||||
} else {
|
||||
logger.error({err: error, url: req.url}, "server error");
|
||||
return res.sendStatus((error != null ? error.statusCode : undefined) || 500);
|
||||
}
|
||||
});
|
||||
|
||||
net = require "net"
|
||||
os = require "os"
|
||||
const net = require("net");
|
||||
const os = require("os");
|
||||
|
||||
STATE = "up"
|
||||
let STATE = "up";
|
||||
|
||||
|
||||
loadTcpServer = net.createServer (socket) ->
|
||||
socket.on "error", (err)->
|
||||
if err.code == "ECONNRESET"
|
||||
# this always comes up, we don't know why
|
||||
return
|
||||
logger.err err:err, "error with socket on load check"
|
||||
socket.destroy()
|
||||
const loadTcpServer = net.createServer(function(socket) {
|
||||
socket.on("error", function(err){
|
||||
if (err.code === "ECONNRESET") {
|
||||
// this always comes up, we don't know why
|
||||
return;
|
||||
}
|
||||
logger.err({err}, "error with socket on load check");
|
||||
return socket.destroy();
|
||||
});
|
||||
|
||||
if STATE == "up" and Settings.internal.load_balancer_agent.report_load
|
||||
currentLoad = os.loadavg()[0]
|
||||
if ((STATE === "up") && Settings.internal.load_balancer_agent.report_load) {
|
||||
let availableWorkingCpus;
|
||||
const currentLoad = os.loadavg()[0];
|
||||
|
||||
# staging clis's have 1 cpu core only
|
||||
if os.cpus().length == 1
|
||||
availableWorkingCpus = 1
|
||||
else
|
||||
availableWorkingCpus = os.cpus().length - 1
|
||||
// staging clis's have 1 cpu core only
|
||||
if (os.cpus().length === 1) {
|
||||
availableWorkingCpus = 1;
|
||||
} else {
|
||||
availableWorkingCpus = os.cpus().length - 1;
|
||||
}
|
||||
|
||||
freeLoad = availableWorkingCpus - currentLoad
|
||||
freeLoadPercentage = Math.round((freeLoad / availableWorkingCpus) * 100)
|
||||
if freeLoadPercentage <= 0
|
||||
freeLoadPercentage = 1 # when its 0 the server is set to drain and will move projects to different servers
|
||||
socket.write("up, #{freeLoadPercentage}%\n", "ASCII")
|
||||
socket.end()
|
||||
else
|
||||
socket.write("#{STATE}\n", "ASCII")
|
||||
socket.end()
|
||||
const freeLoad = availableWorkingCpus - currentLoad;
|
||||
let freeLoadPercentage = Math.round((freeLoad / availableWorkingCpus) * 100);
|
||||
if (freeLoadPercentage <= 0) {
|
||||
freeLoadPercentage = 1; // when its 0 the server is set to drain and will move projects to different servers
|
||||
}
|
||||
socket.write(`up, ${freeLoadPercentage}%\n`, "ASCII");
|
||||
return socket.end();
|
||||
} else {
|
||||
socket.write(`${STATE}\n`, "ASCII");
|
||||
return socket.end();
|
||||
}
|
||||
});
|
||||
|
||||
loadHttpServer = express()
|
||||
const loadHttpServer = express();
|
||||
|
||||
loadHttpServer.post "/state/up", (req, res, next) ->
|
||||
STATE = "up"
|
||||
logger.info "getting message to set server to down"
|
||||
res.sendStatus 204
|
||||
loadHttpServer.post("/state/up", function(req, res, next) {
|
||||
STATE = "up";
|
||||
logger.info("getting message to set server to down");
|
||||
return res.sendStatus(204);
|
||||
});
|
||||
|
||||
loadHttpServer.post "/state/down", (req, res, next) ->
|
||||
STATE = "down"
|
||||
logger.info "getting message to set server to down"
|
||||
res.sendStatus 204
|
||||
loadHttpServer.post("/state/down", function(req, res, next) {
|
||||
STATE = "down";
|
||||
logger.info("getting message to set server to down");
|
||||
return res.sendStatus(204);
|
||||
});
|
||||
|
||||
loadHttpServer.post "/state/maint", (req, res, next) ->
|
||||
STATE = "maint"
|
||||
logger.info "getting message to set server to maint"
|
||||
res.sendStatus 204
|
||||
loadHttpServer.post("/state/maint", function(req, res, next) {
|
||||
STATE = "maint";
|
||||
logger.info("getting message to set server to maint");
|
||||
return res.sendStatus(204);
|
||||
});
|
||||
|
||||
|
||||
port = (Settings.internal?.clsi?.port or 3013)
|
||||
host = (Settings.internal?.clsi?.host or "localhost")
|
||||
const port = (__guard__(Settings.internal != null ? Settings.internal.clsi : undefined, x => x.port) || 3013);
|
||||
const host = (__guard__(Settings.internal != null ? Settings.internal.clsi : undefined, x1 => x1.host) || "localhost");
|
||||
|
||||
load_tcp_port = Settings.internal.load_balancer_agent.load_port
|
||||
load_http_port = Settings.internal.load_balancer_agent.local_port
|
||||
const load_tcp_port = Settings.internal.load_balancer_agent.load_port;
|
||||
const load_http_port = Settings.internal.load_balancer_agent.local_port;
|
||||
|
||||
if !module.parent # Called directly
|
||||
app.listen port, host, (error) ->
|
||||
logger.info "CLSI starting up, listening on #{host}:#{port}"
|
||||
if (!module.parent) { // Called directly
|
||||
app.listen(port, host, error => logger.info(`CLSI starting up, listening on ${host}:${port}`));
|
||||
|
||||
loadTcpServer.listen load_tcp_port, host, (error) ->
|
||||
throw error if error?
|
||||
logger.info "Load tcp agent listening on load port #{load_tcp_port}"
|
||||
loadTcpServer.listen(load_tcp_port, host, function(error) {
|
||||
if (error != null) { throw error; }
|
||||
return logger.info(`Load tcp agent listening on load port ${load_tcp_port}`);
|
||||
});
|
||||
|
||||
loadHttpServer.listen load_http_port, host, (error) ->
|
||||
throw error if error?
|
||||
logger.info "Load http agent listening on load port #{load_http_port}"
|
||||
loadHttpServer.listen(load_http_port, host, function(error) {
|
||||
if (error != null) { throw error; }
|
||||
return logger.info(`Load http agent listening on load port ${load_http_port}`);
|
||||
});
|
||||
}
|
||||
|
||||
module.exports = app
|
||||
module.exports = app;
|
||||
|
||||
setInterval () ->
|
||||
ProjectPersistenceManager.clearExpiredProjects()
|
||||
, tenMinutes = 10 * 60 * 1000
|
||||
setInterval(() => ProjectPersistenceManager.clearExpiredProjects()
|
||||
, (tenMinutes = 10 * 60 * 1000));
|
||||
|
||||
|
||||
function __guard__(value, transform) {
|
||||
return (typeof value !== 'undefined' && value !== null) ? transform(value) : undefined;
|
||||
}
|
|
@ -1,71 +1,89 @@
|
|||
Path = require "path"
|
||||
const Path = require("path");
|
||||
|
||||
module.exports =
|
||||
# Options are passed to Sequelize.
|
||||
# See http://sequelizejs.com/documentation#usage-options for details
|
||||
mysql:
|
||||
clsi:
|
||||
database: "clsi"
|
||||
username: "clsi"
|
||||
dialect: "sqlite"
|
||||
storage: process.env["SQLITE_PATH"] or Path.resolve(__dirname + "/../db.sqlite")
|
||||
pool:
|
||||
max: 1
|
||||
module.exports = {
|
||||
// Options are passed to Sequelize.
|
||||
// See http://sequelizejs.com/documentation#usage-options for details
|
||||
mysql: {
|
||||
clsi: {
|
||||
database: "clsi",
|
||||
username: "clsi",
|
||||
dialect: "sqlite",
|
||||
storage: process.env["SQLITE_PATH"] || Path.resolve(__dirname + "/../db.sqlite"),
|
||||
pool: {
|
||||
max: 1,
|
||||
min: 1
|
||||
retry:
|
||||
},
|
||||
retry: {
|
||||
max: 10
|
||||
}
|
||||
}
|
||||
},
|
||||
|
||||
compileSizeLimit: process.env["COMPILE_SIZE_LIMIT"] or "7mb"
|
||||
compileSizeLimit: process.env["COMPILE_SIZE_LIMIT"] || "7mb",
|
||||
|
||||
path:
|
||||
compilesDir: Path.resolve(__dirname + "/../compiles")
|
||||
clsiCacheDir: Path.resolve(__dirname + "/../cache")
|
||||
synctexBaseDir: (project_id) -> Path.join(@compilesDir, project_id)
|
||||
path: {
|
||||
compilesDir: Path.resolve(__dirname + "/../compiles"),
|
||||
clsiCacheDir: Path.resolve(__dirname + "/../cache"),
|
||||
synctexBaseDir(project_id) { return Path.join(this.compilesDir, project_id); }
|
||||
},
|
||||
|
||||
internal:
|
||||
clsi:
|
||||
port: 3013
|
||||
host: process.env["LISTEN_ADDRESS"] or "localhost"
|
||||
internal: {
|
||||
clsi: {
|
||||
port: 3013,
|
||||
host: process.env["LISTEN_ADDRESS"] || "localhost"
|
||||
},
|
||||
|
||||
load_balancer_agent:
|
||||
report_load:true
|
||||
load_port: 3048
|
||||
load_balancer_agent: {
|
||||
report_load:true,
|
||||
load_port: 3048,
|
||||
local_port: 3049
|
||||
apis:
|
||||
clsi:
|
||||
url: "http://#{process.env['CLSI_HOST'] or 'localhost'}:3013"
|
||||
}
|
||||
},
|
||||
apis: {
|
||||
clsi: {
|
||||
url: `http://${process.env['CLSI_HOST'] || 'localhost'}:3013`
|
||||
}
|
||||
},
|
||||
|
||||
|
||||
smokeTest: process.env["SMOKE_TEST"] or false
|
||||
project_cache_length_ms: 1000 * 60 * 60 * 24
|
||||
parallelFileDownloads: process.env["FILESTORE_PARALLEL_FILE_DOWNLOADS"] or 1
|
||||
parallelSqlQueryLimit: process.env["FILESTORE_PARALLEL_SQL_QUERY_LIMIT"] or 1
|
||||
filestoreDomainOveride: process.env["FILESTORE_DOMAIN_OVERRIDE"]
|
||||
texliveImageNameOveride: process.env["TEX_LIVE_IMAGE_NAME_OVERRIDE"]
|
||||
sentry:
|
||||
smokeTest: process.env["SMOKE_TEST"] || false,
|
||||
project_cache_length_ms: 1000 * 60 * 60 * 24,
|
||||
parallelFileDownloads: process.env["FILESTORE_PARALLEL_FILE_DOWNLOADS"] || 1,
|
||||
parallelSqlQueryLimit: process.env["FILESTORE_PARALLEL_SQL_QUERY_LIMIT"] || 1,
|
||||
filestoreDomainOveride: process.env["FILESTORE_DOMAIN_OVERRIDE"],
|
||||
texliveImageNameOveride: process.env["TEX_LIVE_IMAGE_NAME_OVERRIDE"],
|
||||
sentry: {
|
||||
dsn: process.env['SENTRY_DSN']
|
||||
}
|
||||
};
|
||||
|
||||
|
||||
if process.env["DOCKER_RUNNER"]
|
||||
module.exports.clsi =
|
||||
dockerRunner: process.env["DOCKER_RUNNER"] == "true"
|
||||
docker:
|
||||
image: process.env["TEXLIVE_IMAGE"] or "quay.io/sharelatex/texlive-full:2017.1"
|
||||
env:
|
||||
if (process.env["DOCKER_RUNNER"]) {
|
||||
let seccomp_profile_path;
|
||||
module.exports.clsi = {
|
||||
dockerRunner: process.env["DOCKER_RUNNER"] === "true",
|
||||
docker: {
|
||||
image: process.env["TEXLIVE_IMAGE"] || "quay.io/sharelatex/texlive-full:2017.1",
|
||||
env: {
|
||||
HOME: "/tmp"
|
||||
socketPath: "/var/run/docker.sock"
|
||||
user: process.env["TEXLIVE_IMAGE_USER"] or "tex"
|
||||
expireProjectAfterIdleMs: 24 * 60 * 60 * 1000
|
||||
},
|
||||
socketPath: "/var/run/docker.sock",
|
||||
user: process.env["TEXLIVE_IMAGE_USER"] || "tex"
|
||||
},
|
||||
expireProjectAfterIdleMs: 24 * 60 * 60 * 1000,
|
||||
checkProjectsIntervalMs: 10 * 60 * 1000
|
||||
};
|
||||
|
||||
try
|
||||
seccomp_profile_path = Path.resolve(__dirname + "/../seccomp/clsi-profile.json")
|
||||
module.exports.clsi.docker.seccomp_profile = JSON.stringify(JSON.parse(require("fs").readFileSync(seccomp_profile_path)))
|
||||
catch error
|
||||
console.log error, "could not load seccom profile from #{seccomp_profile_path}"
|
||||
try {
|
||||
seccomp_profile_path = Path.resolve(__dirname + "/../seccomp/clsi-profile.json");
|
||||
module.exports.clsi.docker.seccomp_profile = JSON.stringify(JSON.parse(require("fs").readFileSync(seccomp_profile_path)));
|
||||
} catch (error) {
|
||||
console.log(error, `could not load seccom profile from ${seccomp_profile_path}`);
|
||||
}
|
||||
|
||||
module.exports.path.synctexBaseDir = -> "/compile"
|
||||
module.exports.path.synctexBaseDir = () => "/compile";
|
||||
|
||||
module.exports.path.sandboxedCompilesHostDir = process.env["COMPILES_HOST_DIR"]
|
||||
module.exports.path.sandboxedCompilesHostDir = process.env["COMPILES_HOST_DIR"];
|
||||
|
||||
module.exports.path.synctexBinHostPath = process.env["SYNCTEX_BIN_HOST_PATH"]
|
||||
module.exports.path.synctexBinHostPath = process.env["SYNCTEX_BIN_HOST_PATH"];
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue