[misc] run format:fix

This commit is contained in:
Jakob Ackermann 2021-07-07 12:43:33 +01:00
parent c0dd3df094
commit 1b63336d36
No known key found for this signature in database
GPG key ID: 30C56800FCA3828A
3 changed files with 993 additions and 839 deletions

View file

@ -16,42 +16,36 @@
* DS207: Consider shorter variations of null checks
* Full docs: https://github.com/decaffeinate/decaffeinate/blob/master/docs/suggestions.md
*/
const coffee = require("coffee-script");
const fs = require("fs");
const {
spawn
} = require("child_process");
const {
exec
} = require("child_process");
const rimraf = require("rimraf");
const Path = require("path");
const semver = require("semver");
const knox = require("knox");
const crypto = require("crypto");
const async = require("async");
const settings = require("settings-sharelatex");
const _ = require("underscore");
const coffee = require('coffee-script')
const fs = require('fs')
const { spawn } = require('child_process')
const { exec } = require('child_process')
const rimraf = require('rimraf')
const Path = require('path')
const semver = require('semver')
const knox = require('knox')
const crypto = require('crypto')
const async = require('async')
const settings = require('settings-sharelatex')
const _ = require('underscore')
const SERVICES = require("./config/services");
const SERVICES = require('./config/services')
module.exports = function (grunt) {
let Helpers;
let service;
grunt.loadNpmTasks('grunt-bunyan');
grunt.loadNpmTasks('grunt-execute');
grunt.loadNpmTasks('grunt-available-tasks');
grunt.loadNpmTasks('grunt-concurrent');
grunt.loadNpmTasks("grunt-contrib-coffee");
grunt.loadNpmTasks("grunt-shell");
let Helpers
let service
grunt.loadNpmTasks('grunt-bunyan')
grunt.loadNpmTasks('grunt-execute')
grunt.loadNpmTasks('grunt-available-tasks')
grunt.loadNpmTasks('grunt-concurrent')
grunt.loadNpmTasks('grunt-contrib-coffee')
grunt.loadNpmTasks('grunt-shell')
grunt.task.loadTasks("./tasks");
grunt.task.loadTasks('./tasks')
const execute = {};
const execute = {}
for (service of Array.from(SERVICES)) {
execute[service.name] =
{src: `${service.name}/app.js`};
execute[service.name] = { src: `${service.name}/app.js` }
}
grunt.initConfig({
@ -59,178 +53,221 @@ module.exports = function(grunt) {
concurrent: {
all: {
tasks: (((() => {
const result = [];
for (service of Array.from(SERVICES)) { result.push(`run:${service.name}`);
tasks: (() => {
const result = []
for (service of Array.from(SERVICES)) {
result.push(`run:${service.name}`)
}
return result;
})())),
return result
})(),
options: {
limit: SERVICES.length,
logConcurrentOutput: true
logConcurrentOutput: true,
},
},
},
availabletasks: {
tasks: {
options: {
filter: 'exclude',
tasks: [
'concurrent',
'execute',
'bunyan',
'availabletasks'
],
tasks: ['concurrent', 'execute', 'bunyan', 'availabletasks'],
groups: {
"Run tasks": [
"run",
"run:all",
"default"
].concat(((() => {
const result1 = [];
for (service of Array.from(SERVICES)) { result1.push(`run:${service.name}`);
'Run tasks': ['run', 'run:all', 'default'].concat(
(() => {
const result1 = []
for (service of Array.from(SERVICES)) {
result1.push(`run:${service.name}`)
}
return result1;
})())),
"Misc": [
"help"
return result1
})()
),
Misc: ['help'],
'Install tasks': (() => {
const result2 = []
for (service of Array.from(SERVICES)) {
result2.push(`install:${service.name}`)
}
return result2
})().concat(['install:all', 'install']),
'Update tasks': (() => {
const result3 = []
for (service of Array.from(SERVICES)) {
result3.push(`update:${service.name}`)
}
return result3
})().concat(['update:all', 'update']),
Checks: [
'check',
'check:redis',
'check:latexmk',
'check:s3',
'check:make',
'check:mongo',
],
"Install tasks": ((() => {
const result2 = [];
for (service of Array.from(SERVICES)) { result2.push(`install:${service.name}`);
}
return result2;
})()).concat(["install:all", "install"]),
"Update tasks": ((() => {
const result3 = [];
for (service of Array.from(SERVICES)) { result3.push(`update:${service.name}`);
}
return result3;
})()).concat(["update:all", "update"]),
"Checks": ["check", "check:redis", "check:latexmk", "check:s3", "check:make", "check:mongo"]
}
}
}
}});
},
},
},
},
})
for (service of Array.from(SERVICES)) {
((service => grunt.registerTask(`install:${service.name}`, `Download and set up the ${service.name} service`, function() {
const done = this.async();
return Helpers.installService(service, done);
})))(service);
;(service =>
grunt.registerTask(
`install:${service.name}`,
`Download and set up the ${service.name} service`,
function () {
const done = this.async()
return Helpers.installService(service, done)
}
))(service)
}
grunt.registerTask('install:all', "Download and set up all ShareLaTeX services",
[].concat(
((() => {
const result4 = [];
for (service of Array.from(SERVICES)) { result4.push(`install:${service.name}`);
grunt.registerTask(
'install:all',
'Download and set up all ShareLaTeX services',
[]
.concat(
(() => {
const result4 = []
for (service of Array.from(SERVICES)) {
result4.push(`install:${service.name}`)
}
return result4;
})())
).concat(['postinstall'])
);
grunt.registerTask('install', 'install:all');
grunt.registerTask('postinstall', 'Explain postinstall steps', function() {
return Helpers.postinstallMessage(this.async());
});
grunt.registerTask('update:all', "Checkout and update all ShareLaTeX services",
["check:make"].concat(
((() => {
const result5 = [];
for (service of Array.from(SERVICES)) { result5.push(`update:${service.name}`);
}
return result5;
})())
return result4
})()
)
.concat(['postinstall'])
)
);
grunt.registerTask('update', 'update:all');
grunt.registerTask('run', "Run all of the sharelatex processes", ['concurrent:all']);
grunt.registerTask('run:all', 'run');
grunt.registerTask('help', 'Display this help list', 'availabletasks');
grunt.registerTask('default', 'run');
grunt.registerTask('install', 'install:all')
grunt.registerTask('postinstall', 'Explain postinstall steps', function () {
return Helpers.postinstallMessage(this.async())
})
grunt.registerTask("check:redis", "Check that redis is installed and running", function() {
return Helpers.checkRedisConnect(this.async());
});
grunt.registerTask(
'update:all',
'Checkout and update all ShareLaTeX services',
['check:make'].concat(
(() => {
const result5 = []
for (service of Array.from(SERVICES)) {
result5.push(`update:${service.name}`)
}
return result5
})()
)
)
grunt.registerTask('update', 'update:all')
grunt.registerTask('run', 'Run all of the sharelatex processes', [
'concurrent:all',
])
grunt.registerTask('run:all', 'run')
grunt.registerTask("check:mongo", "Check that mongo is installed", function() {
return Helpers.checkMongoConnect(this.async());
});
grunt.registerTask('help', 'Display this help list', 'availabletasks')
grunt.registerTask('default', 'run')
grunt.registerTask("check", "Check that you have the required dependencies installed", ["check:redis", "check:mongo", "check:make"]);
grunt.registerTask(
'check:redis',
'Check that redis is installed and running',
function () {
return Helpers.checkRedisConnect(this.async())
}
)
grunt.registerTask("check:make", "Check that make is installed", function() {
return Helpers.checkMake(this.async());
});
grunt.registerTask(
'check:mongo',
'Check that mongo is installed',
function () {
return Helpers.checkMongoConnect(this.async())
}
)
grunt.registerTask(
'check',
'Check that you have the required dependencies installed',
['check:redis', 'check:mongo', 'check:make']
)
return Helpers = {
grunt.registerTask('check:make', 'Check that make is installed', function () {
return Helpers.checkMake(this.async())
})
return (Helpers = {
installService(service, callback) {
if (callback == null) { callback = function(error) {}; }
console.log(`Installing ${service.name}`);
if (callback == null) {
callback = function (error) {}
}
console.log(`Installing ${service.name}`)
return Helpers.cloneGitRepo(service, function (error) {
if (error != null) {
return callback(error);
return callback(error)
} else {
return callback();
return callback()
}
});
})
},
cloneGitRepo(service, callback) {
if (callback == null) { callback = function(error) {}; }
const repo_src = service.repo;
const dir = service.name;
if (callback == null) {
callback = function (error) {}
}
const repo_src = service.repo
const dir = service.name
if (!fs.existsSync(dir)) {
const proc = spawn("git", [
"clone",
repo_src,
dir
], {stdio: "inherit"});
return proc.on("close", () => Helpers.checkoutVersion(service, callback));
const proc = spawn('git', ['clone', repo_src, dir], {
stdio: 'inherit',
})
return proc.on('close', () =>
Helpers.checkoutVersion(service, callback)
)
} else {
console.log(`${dir} already installed, skipping.`);
return callback();
console.log(`${dir} already installed, skipping.`)
return callback()
}
},
checkoutVersion(service, callback) {
if (callback == null) { callback = function(error) {}; }
const dir = service.name;
grunt.log.write(`checking out ${service.name} ${service.version}`);
const proc = spawn("git", ["checkout", service.version], {stdio: "inherit", cwd: dir});
return proc.on("close", () => callback());
if (callback == null) {
callback = function (error) {}
}
const dir = service.name
grunt.log.write(`checking out ${service.name} ${service.version}`)
const proc = spawn('git', ['checkout', service.version], {
stdio: 'inherit',
cwd: dir,
})
return proc.on('close', () => callback())
},
postinstallMessage(callback) {
if (callback == null) { callback = function(error) {}; }
if (callback == null) {
callback = function (error) {}
}
grunt.log.write(`\
Services cloned:
${(() => {
const result6 = [];
for (service of Array.from(SERVICES)) { result6.push(service.name);
const result6 = []
for (service of Array.from(SERVICES)) {
result6.push(service.name)
}
return result6;
return result6
})()}
To install services run:
$ source bin/install-services
This will install the required node versions and run \`npm install\` for each service.
See https://github.com/sharelatex/sharelatex/pull/549 for more info.\
`
);
return callback();
`)
return callback()
},
checkMake(callback) {
if (callback == null) { callback = function(error) {}; }
grunt.log.write("Checking make is installed... ");
return exec("make --version", function(error, stdout, stderr) {
if ((error != null) && error.message.match("not found")) {
grunt.log.error("FAIL.");
if (callback == null) {
callback = function (error) {}
}
grunt.log.write('Checking make is installed... ')
return exec('make --version', function (error, stdout, stderr) {
if (error != null && error.message.match('not found')) {
grunt.log.error('FAIL.')
grunt.log.errorlns(`\
Either make is not installed or is not in your path.
@ -238,31 +275,32 @@ On Ubuntu you can install make with:
sudo apt-get install build-essential
\
`
);
return callback(error);
`)
return callback(error)
} else if (error != null) {
return callback(error);
return callback(error)
} else {
grunt.log.write("OK.");
return callback();
grunt.log.write('OK.')
return callback()
}
});
})
},
checkMongoConnect(callback) {
if (callback == null) { callback = function(error) {}; }
grunt.log.write("Checking can connect to mongo");
const mongojs = require("mongojs");
const db = mongojs(settings.mongo.url, ["tags"]);
if (callback == null) {
callback = function (error) {}
}
grunt.log.write('Checking can connect to mongo')
const mongojs = require('mongojs')
const db = mongojs(settings.mongo.url, ['tags'])
db.runCommand({ ping: 1 }, function (err, res) {
if (!err && res.ok) {
grunt.log.write("OK.");
grunt.log.write('OK.')
}
return callback();
});
return callback()
})
return db.on('error', function (err) {
err = "Can not connect to mongodb";
grunt.log.error("FAIL.");
err = 'Can not connect to mongodb'
grunt.log.error('FAIL.')
grunt.log.errorlns(`\
!!!!!!!!!!!!!! MONGO ERROR !!!!!!!!!!!!!!
@ -271,29 +309,30 @@ ShareLaTeX can not talk to the mongodb instance
Check the mongodb instance is running and accessible on env var SHARELATEX_MONGO_URL
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\
`
);
throw new Error("Can not connect to Mongodb");
return callback(err);
});
`)
throw new Error('Can not connect to Mongodb')
return callback(err)
})
},
checkRedisConnect(callback) {
if (callback == null) { callback = function(error) {}; }
grunt.log.write("Checking can connect to redis\n");
const rclient = require("redis").createClient(settings.redis.web);
if (callback == null) {
callback = function (error) {}
}
grunt.log.write('Checking can connect to redis\n')
const rclient = require('redis').createClient(settings.redis.web)
rclient.ping(function (err, res) {
if ((err == null)) {
grunt.log.write("OK.");
if (err == null) {
grunt.log.write('OK.')
} else {
throw new Error("Can not connect to redis");
throw new Error('Can not connect to redis')
}
return callback();
});
return callback()
})
const errorHandler = _.once(function (err) {
err = "Can not connect to redis";
grunt.log.error("FAIL.");
err = 'Can not connect to redis'
grunt.log.error('FAIL.')
grunt.log.errorlns(`\
!!!!!!!!!!!!!! REDIS ERROR !!!!!!!!!!!!!!
@ -302,19 +341,17 @@ ShareLaTeX can not talk to the redis instance
Check the redis instance is running and accessible on env var SHARELATEX_REDIS_HOST
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\
`
);
throw new Error("Can not connect to redis");
return callback(err);
});
return rclient.on('error', errorHandler);
`)
throw new Error('Can not connect to redis')
return callback(err)
})
return rclient.on('error', errorHandler)
},
})
}
};
};
function __guard__(value, transform) {
return (typeof value !== 'undefined' && value !== null) ? transform(value) : undefined;
return typeof value !== 'undefined' && value !== null
? transform(value)
: undefined
}

View file

@ -12,46 +12,45 @@
* DS207: Consider shorter variations of null checks
* Full docs: https://github.com/decaffeinate/decaffeinate/blob/master/docs/suggestions.md
*/
let allTexLiveDockerImageNames, allTexLiveDockerImages, redisConfig, siteUrl;
let e;
const Path = require('path');
let allTexLiveDockerImageNames, allTexLiveDockerImages, redisConfig, siteUrl
let e
const Path = require('path')
// These credentials are used for authenticating api requests
// between services that may need to go over public channels
const httpAuthUser = "sharelatex";
const httpAuthPass = process.env.WEB_API_PASSWORD;
const httpAuthUsers = {};
httpAuthUsers[httpAuthUser] = httpAuthPass;
const httpAuthUser = 'sharelatex'
const httpAuthPass = process.env.WEB_API_PASSWORD
const httpAuthUsers = {}
httpAuthUsers[httpAuthUser] = httpAuthPass
const parse = function (option) {
if (option != null) {
try {
const opt = JSON.parse(option);
return opt;
const opt = JSON.parse(option)
return opt
} catch (err) {
throw new Error(`problem parsing ${option}, invalid JSON`);
throw new Error(`problem parsing ${option}, invalid JSON`)
}
}
}
};
const parseIntOrFail = function (value) {
const parsedValue = parseInt(value, 10);
const parsedValue = parseInt(value, 10)
if (isNaN(parsedValue)) {
throw new Error(`'${value}' is an invalid integer`);
throw new Error(`'${value}' is an invalid integer`)
}
return parsedValue
}
return parsedValue;
};
const DATA_DIR = '/var/lib/sharelatex/data';
const TMP_DIR = '/var/lib/sharelatex/tmp';
const DATA_DIR = '/var/lib/sharelatex/data'
const TMP_DIR = '/var/lib/sharelatex/tmp'
const settings = {
clsi: {
optimiseInDocker: process.env.OPTIMISE_PDF === 'true'
optimiseInDocker: process.env.OPTIMISE_PDF === 'true',
},
brandPrefix: "",
brandPrefix: '',
allowAnonymousReadAndWriteSharing:
process.env.SHARELATEX_ALLOW_ANONYMOUS_READ_AND_WRITE_SHARING === 'true',
@ -66,7 +65,7 @@ const settings = {
//
// The following works out of the box with Mongo's default settings:
mongo: {
url : process.env.SHARELATEX_MONGO_URL || 'mongodb://dockerhost/sharelatex'
url: process.env.SHARELATEX_MONGO_URL || 'mongodb://dockerhost/sharelatex',
},
// Redis is used in ShareLaTeX for high volume queries, like real-time
@ -75,36 +74,72 @@ const settings = {
// The following config will work with Redis's default settings:
redis: {
web: (redisConfig = {
host: process.env.SHARELATEX_REDIS_HOST || "dockerhost",
port: process.env.SHARELATEX_REDIS_PORT || "6379",
host: process.env.SHARELATEX_REDIS_HOST || 'dockerhost',
port: process.env.SHARELATEX_REDIS_PORT || '6379',
password: process.env.SHARELATEX_REDIS_PASS || undefined,
key_schema: {
// document-updater
blockingKey({doc_id}) { return `Blocking:${doc_id}`; },
docLines({doc_id}) { return `doclines:${doc_id}`; },
docOps({doc_id}) { return `DocOps:${doc_id}`; },
docVersion({doc_id}) { return `DocVersion:${doc_id}`; },
docHash({doc_id}) { return `DocHash:${doc_id}`; },
projectKey({doc_id}) { return `ProjectId:${doc_id}`; },
docsInProject({project_id}) { return `DocsIn:${project_id}`; },
ranges({doc_id}) { return `Ranges:${doc_id}`; },
blockingKey({ doc_id }) {
return `Blocking:${doc_id}`
},
docLines({ doc_id }) {
return `doclines:${doc_id}`
},
docOps({ doc_id }) {
return `DocOps:${doc_id}`
},
docVersion({ doc_id }) {
return `DocVersion:${doc_id}`
},
docHash({ doc_id }) {
return `DocHash:${doc_id}`
},
projectKey({ doc_id }) {
return `ProjectId:${doc_id}`
},
docsInProject({ project_id }) {
return `DocsIn:${project_id}`
},
ranges({ doc_id }) {
return `Ranges:${doc_id}`
},
// document-updater:realtime
pendingUpdates({doc_id}) { return `PendingUpdates:${doc_id}`; },
pendingUpdates({ doc_id }) {
return `PendingUpdates:${doc_id}`
},
// document-updater:history
uncompressedHistoryOps({doc_id}) { return `UncompressedHistoryOps:${doc_id}`; },
docsWithHistoryOps({project_id}) { return `DocsWithHistoryOps:${project_id}`; },
uncompressedHistoryOps({ doc_id }) {
return `UncompressedHistoryOps:${doc_id}`
},
docsWithHistoryOps({ project_id }) {
return `DocsWithHistoryOps:${project_id}`
},
// document-updater:lock
blockingKey({doc_id}) { return `Blocking:${doc_id}`; },
blockingKey({ doc_id }) {
return `Blocking:${doc_id}`
},
// track-changes:lock
historyLock({doc_id}) { return `HistoryLock:${doc_id}`; },
historyIndexLock({project_id}) { return `HistoryIndexLock:${project_id}`; },
historyLock({ doc_id }) {
return `HistoryLock:${doc_id}`
},
historyIndexLock({ project_id }) {
return `HistoryIndexLock:${project_id}`
},
// track-changes:history
uncompressedHistoryOps({doc_id}) { return `UncompressedHistoryOps:${doc_id}`; },
docsWithHistoryOps({project_id}) { return `DocsWithHistoryOps:${project_id}`; },
uncompressedHistoryOps({ doc_id }) {
return `UncompressedHistoryOps:${doc_id}`
},
docsWithHistoryOps({ project_id }) {
return `DocsWithHistoryOps:${project_id}`
},
// realtime
clientsInProject({project_id}) { return `clients_in_project:${project_id}`; },
connectedUser({project_id, client_id}){ return `connected_user:${project_id}:${client_id}`; }
}
clientsInProject({ project_id }) {
return `clients_in_project:${project_id}`
},
connectedUser({ project_id, client_id }) {
return `connected_user:${project_id}:${client_id}`
},
},
}),
fairy: redisConfig,
// track-changes and document-updater
@ -115,7 +150,7 @@ const settings = {
websessions: redisConfig,
api: redisConfig,
pubsub: redisConfig,
project_history: redisConfig
project_history: redisConfig,
},
// The compile server (the clsi) uses a SQL database to cache files and
@ -129,12 +164,12 @@ const settings = {
//
mysql: {
clsi: {
database: "clsi",
username: "clsi",
password: "",
dialect: "sqlite",
storage: Path.join(DATA_DIR, "db.sqlite")
}
database: 'clsi',
username: 'clsi',
password: '',
dialect: 'sqlite',
storage: Path.join(DATA_DIR, 'db.sqlite'),
},
},
// File storage
@ -143,11 +178,11 @@ const settings = {
// ShareLaTeX can store binary files like images either locally or in Amazon
// S3. The default is locally:
filestore: {
backend: "fs",
backend: 'fs',
stores: {
user_files: Path.join(DATA_DIR, "user_files"),
template_files: Path.join(DATA_DIR, "template_files")
}
user_files: Path.join(DATA_DIR, 'user_files'),
template_files: Path.join(DATA_DIR, 'template_files'),
},
},
// To use Amazon S3 as a storage backend, comment out the above config, and
@ -163,7 +198,7 @@ const settings = {
//
trackchanges: {
continueOnError: true
continueOnError: true,
},
// Local disk caching
@ -172,15 +207,15 @@ const settings = {
// If we ever need to write something to disk (e.g. incoming requests
// that need processing but may be too big for memory), then write
// them to disk here:
dumpFolder: Path.join(TMP_DIR, "dumpFolder"),
dumpFolder: Path.join(TMP_DIR, 'dumpFolder'),
// Where to write uploads before they are processed
uploadFolder: Path.join(TMP_DIR, "uploads"),
uploadFolder: Path.join(TMP_DIR, 'uploads'),
// Where to write the project to disk before running LaTeX on it
compilesDir: Path.join(DATA_DIR, "compiles"),
compilesDir: Path.join(DATA_DIR, 'compiles'),
// Where to cache downloaded URLs for the CLSI
clsiCacheDir: Path.join(DATA_DIR, "cache"),
clsiCacheDir: Path.join(DATA_DIR, 'cache'),
// Where to write the output files to disk after running LaTeX
outputDir: Path.join(DATA_DIR, "output")
outputDir: Path.join(DATA_DIR, 'output'),
},
// Server Config
@ -191,23 +226,27 @@ const settings = {
siteUrl: (siteUrl = process.env.SHARELATEX_SITE_URL || 'http://localhost'),
// The name this is used to describe your ShareLaTeX Installation
appName: process.env.SHARELATEX_APP_NAME || "ShareLaTeX (Community Edition)",
appName: process.env.SHARELATEX_APP_NAME || 'ShareLaTeX (Community Edition)',
restrictInvitesToExistingAccounts: process.env.SHARELATEX_RESTRICT_INVITES_TO_EXISTING_ACCOUNTS === 'true',
restrictInvitesToExistingAccounts:
process.env.SHARELATEX_RESTRICT_INVITES_TO_EXISTING_ACCOUNTS === 'true',
nav: {
title: process.env.SHARELATEX_NAV_TITLE || process.env.SHARELATEX_APP_NAME || "ShareLaTeX Community Edition"
title:
process.env.SHARELATEX_NAV_TITLE ||
process.env.SHARELATEX_APP_NAME ||
'ShareLaTeX Community Edition',
},
// The email address which users will be directed to as the main point of
// contact for this installation of ShareLaTeX.
adminEmail: process.env.SHARELATEX_ADMIN_EMAIL || "placeholder@example.com",
adminEmail: process.env.SHARELATEX_ADMIN_EMAIL || 'placeholder@example.com',
// If provided, a sessionSecret is used to sign cookies so that they cannot be
// spoofed. This is recommended.
security: {
sessionSecret: process.env.SHARELATEX_SESSION_SECRET || process.env.CRYPTO_RANDOM
sessionSecret:
process.env.SHARELATEX_SESSION_SECRET || process.env.CRYPTO_RANDOM,
},
// These credentials are used for authenticating api requests
@ -224,7 +263,7 @@ const settings = {
// If you are running ShareLaTeX over https, set this to true to send the
// cookie with a secure flag (recommended).
secureCookie: (process.env.SHARELATEX_SECURE_COOKIE != null),
secureCookie: process.env.SHARELATEX_SECURE_COOKIE != null,
// If you are running ShareLaTeX behind a proxy (like Apache, Nginx, etc)
// then set this to true to allow it to correctly detect the forwarded IP
@ -234,22 +273,25 @@ const settings = {
i18n: {
subdomainLang: {
www: {lngCode:process.env.SHARELATEX_SITE_LANGUAGE || "en", url: siteUrl}
www: {
lngCode: process.env.SHARELATEX_SITE_LANGUAGE || 'en',
url: siteUrl,
},
defaultLng: process.env.SHARELATEX_SITE_LANGUAGE || "en"
},
defaultLng: process.env.SHARELATEX_SITE_LANGUAGE || 'en',
},
currentImageName: process.env.TEX_LIVE_DOCKER_IMAGE,
apis: {
web: {
url: "http://localhost:3000",
url: 'http://localhost:3000',
user: httpAuthUser,
pass: httpAuthPass
pass: httpAuthPass,
},
project_history: {
enabled: false
}
enabled: false,
},
},
references: {},
notifications: undefined,
@ -259,36 +301,36 @@ const settings = {
dropbox: true,
versioning: true,
compileTimeout: parseIntOrFail(process.env.COMPILE_TIMEOUT || 180),
compileGroup: "standard",
compileGroup: 'standard',
trackChanges: true,
templates: true,
references: true
references: true,
},
}
};
// # OPTIONAL CONFIGURABLE SETTINGS
if (process.env.SHARELATEX_LEFT_FOOTER != null) {
try {
settings.nav.left_footer = JSON.parse(process.env.SHARELATEX_LEFT_FOOTER);
settings.nav.left_footer = JSON.parse(process.env.SHARELATEX_LEFT_FOOTER)
} catch (error) {
e = error;
console.error("could not parse SHARELATEX_LEFT_FOOTER, not valid JSON");
e = error
console.error('could not parse SHARELATEX_LEFT_FOOTER, not valid JSON')
}
}
if (process.env.SHARELATEX_RIGHT_FOOTER != null) {
settings.nav.right_footer = process.env.SHARELATEX_RIGHT_FOOTER;
settings.nav.right_footer = process.env.SHARELATEX_RIGHT_FOOTER
try {
settings.nav.right_footer = JSON.parse(process.env.SHARELATEX_RIGHT_FOOTER);
settings.nav.right_footer = JSON.parse(process.env.SHARELATEX_RIGHT_FOOTER)
} catch (error1) {
e = error1;
console.error("could not parse SHARELATEX_RIGHT_FOOTER, not valid JSON");
e = error1
console.error('could not parse SHARELATEX_RIGHT_FOOTER, not valid JSON')
}
}
if (process.env.SHARELATEX_HEADER_IMAGE_URL != null) {
settings.nav.custom_logo = process.env.SHARELATEX_HEADER_IMAGE_URL;
settings.nav.custom_logo = process.env.SHARELATEX_HEADER_IMAGE_URL
}
if (process.env.SHARELATEX_HEADER_NAV_LINKS != null) {
@ -299,21 +341,20 @@ if (process.env.SHARELATEX_HEADER_NAV_LINKS != null) {
# See https://github.com/sharelatex/sharelatex/wiki/Configuring-Headers,-Footers-&-Logo
#
# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # #\
`
);
`)
}
if (process.env.SHARELATEX_HEADER_EXTRAS != null) {
try {
settings.nav.header_extras = JSON.parse(process.env.SHARELATEX_HEADER_EXTRAS);
settings.nav.header_extras = JSON.parse(
process.env.SHARELATEX_HEADER_EXTRAS
)
} catch (error2) {
e = error2;
console.error("could not parse SHARELATEX_HEADER_EXTRAS, not valid JSON");
e = error2
console.error('could not parse SHARELATEX_HEADER_EXTRAS, not valid JSON')
}
}
// Sending Email
// -------------
//
@ -323,12 +364,10 @@ if (process.env.SHARELATEX_HEADER_EXTRAS != null) {
//
// http://www.nodemailer.com/docs/transports
if (process.env.SHARELATEX_EMAIL_FROM_ADDRESS != null) {
settings.email = {
fromAddress: process.env.SHARELATEX_EMAIL_FROM_ADDRESS,
replyTo: process.env.SHARELATEX_EMAIL_REPLY_TO || "",
replyTo: process.env.SHARELATEX_EMAIL_REPLY_TO || '',
driver: process.env.SHARELATEX_EMAIL_DRIVER,
parameters: {
// AWS Creds
@ -341,65 +380,73 @@ if (process.env.SHARELATEX_EMAIL_FROM_ADDRESS != null) {
secure: parse(process.env.SHARELATEX_EMAIL_SMTP_SECURE),
ignoreTLS: parse(process.env.SHARELATEX_EMAIL_SMTP_IGNORE_TLS),
name: process.env.SHARELATEX_EMAIL_SMTP_NAME,
logger: process.env.SHARELATEX_EMAIL_SMTP_LOGGER === 'true'
logger: process.env.SHARELATEX_EMAIL_SMTP_LOGGER === 'true',
},
textEncoding: process.env.SHARELATEX_EMAIL_TEXT_ENCODING,
template: {
customFooter: process.env.SHARELATEX_CUSTOM_EMAIL_FOOTER
customFooter: process.env.SHARELATEX_CUSTOM_EMAIL_FOOTER,
},
}
};
if (process.env.SHARELATEX_EMAIL_AWS_SES_REGION != null) {
settings.email.parameters.region = process.env.SHARELATEX_EMAIL_AWS_SES_REGION;
settings.email.parameters.region =
process.env.SHARELATEX_EMAIL_AWS_SES_REGION
}
if ((process.env.SHARELATEX_EMAIL_SMTP_USER != null) || (process.env.SHARELATEX_EMAIL_SMTP_PASS != null)) {
if (
process.env.SHARELATEX_EMAIL_SMTP_USER != null ||
process.env.SHARELATEX_EMAIL_SMTP_PASS != null
) {
settings.email.parameters.auth = {
user: process.env.SHARELATEX_EMAIL_SMTP_USER,
pass: process.env.SHARELATEX_EMAIL_SMTP_PASS
};
pass: process.env.SHARELATEX_EMAIL_SMTP_PASS,
}
}
if (process.env.SHARELATEX_EMAIL_SMTP_TLS_REJECT_UNAUTH != null) {
settings.email.parameters.tls =
{rejectUnauthorized: parse(process.env.SHARELATEX_EMAIL_SMTP_TLS_REJECT_UNAUTH)};
settings.email.parameters.tls = {
rejectUnauthorized: parse(
process.env.SHARELATEX_EMAIL_SMTP_TLS_REJECT_UNAUTH
),
}
}
}
// i18n
if (process.env.SHARELATEX_LANG_DOMAIN_MAPPING != null) {
settings.i18n.subdomainLang = parse(process.env.SHARELATEX_LANG_DOMAIN_MAPPING);
settings.i18n.subdomainLang = parse(
process.env.SHARELATEX_LANG_DOMAIN_MAPPING
)
}
// Password Settings
// -----------
// These restrict the passwords users can use when registering
// opts are from http://antelle.github.io/passfield
if (process.env.SHARELATEX_PASSWORD_VALIDATION_PATTERN || process.env.SHARELATEX_PASSWORD_VALIDATION_MIN_LENGTH || process.env.SHARELATEX_PASSWORD_VALIDATION_MAX_LENGTH) {
if (
process.env.SHARELATEX_PASSWORD_VALIDATION_PATTERN ||
process.env.SHARELATEX_PASSWORD_VALIDATION_MIN_LENGTH ||
process.env.SHARELATEX_PASSWORD_VALIDATION_MAX_LENGTH
) {
settings.passwordStrengthOptions = {
pattern: process.env.SHARELATEX_PASSWORD_VALIDATION_PATTERN || "aA$3",
length: {min:process.env.SHARELATEX_PASSWORD_VALIDATION_MIN_LENGTH || 8, max: process.env.SHARELATEX_PASSWORD_VALIDATION_MAX_LENGTH || 150}
};
pattern: process.env.SHARELATEX_PASSWORD_VALIDATION_PATTERN || 'aA$3',
length: {
min: process.env.SHARELATEX_PASSWORD_VALIDATION_MIN_LENGTH || 8,
max: process.env.SHARELATEX_PASSWORD_VALIDATION_MAX_LENGTH || 150,
},
}
}
// ######################
// ShareLaTeX Server Pro
// ######################
if (parse(process.env.SHARELATEX_IS_SERVER_PRO) === true) {
settings.bypassPercentageRollouts = true;
settings.apis.references =
{url: "http://localhost:3040"};
settings.bypassPercentageRollouts = true
settings.apis.references = { url: 'http://localhost:3040' }
}
// LDAP - SERVER PRO ONLY
// ----------
@ -411,18 +458,21 @@ if (process.env.SHARELATEX_LDAP_HOST) {
# See https://github.com/sharelatex/sharelatex/wiki/Server-Pro:-LDAP-Config
#
# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # #\
`
);
`)
}
if (process.env.SHARELATEX_LDAP_URL) {
let _ldap_connect_timeout, _ldap_group_search_attribs, _ldap_search_attribs, _ldap_timeout;
settings.externalAuth = true;
let _ldap_connect_timeout,
_ldap_group_search_attribs,
_ldap_search_attribs,
_ldap_timeout
settings.externalAuth = true
settings.ldap = {
emailAtt: process.env.SHARELATEX_LDAP_EMAIL_ATT,
nameAtt: process.env.SHARELATEX_LDAP_NAME_ATT,
lastNameAtt: process.env.SHARELATEX_LDAP_LAST_NAME_ATT,
updateUserDetailsOnLogin: process.env.SHARELATEX_LDAP_UPDATE_USER_DETAILS_ON_LOGIN === 'true',
updateUserDetailsOnLogin:
process.env.SHARELATEX_LDAP_UPDATE_USER_DETAILS_ON_LOGIN === 'true',
placeholder: process.env.SHARELATEX_LDAP_PLACEHOLDER,
server: {
url: process.env.SHARELATEX_LDAP_URL,
@ -432,94 +482,108 @@ if (process.env.SHARELATEX_LDAP_URL) {
searchBase: process.env.SHARELATEX_LDAP_SEARCH_BASE,
searchScope: process.env.SHARELATEX_LDAP_SEARCH_SCOPE,
searchFilter: process.env.SHARELATEX_LDAP_SEARCH_FILTER,
searchAttributes: (
(_ldap_search_attribs = process.env.SHARELATEX_LDAP_SEARCH_ATTRIBUTES) ?
(() => { try {
return JSON.parse(_ldap_search_attribs);
searchAttributes: (_ldap_search_attribs =
process.env.SHARELATEX_LDAP_SEARCH_ATTRIBUTES)
? (() => {
try {
return JSON.parse(_ldap_search_attribs)
} catch (error3) {
e = error3;
return console.error("could not parse SHARELATEX_LDAP_SEARCH_ATTRIBUTES");
} })()
:
undefined
),
e = error3
return console.error(
'could not parse SHARELATEX_LDAP_SEARCH_ATTRIBUTES'
)
}
})()
: undefined,
groupDnProperty: process.env.SHARELATEX_LDAP_GROUP_DN_PROPERTY,
groupSearchBase: process.env.SHARELATEX_LDAP_GROUP_SEARCH_BASE,
groupSearchScope: process.env.SHARELATEX_LDAP_GROUP_SEARCH_SCOPE,
groupSearchFilter: process.env.SHARELATEX_LDAP_GROUP_SEARCH_FILTER,
groupSearchAttributes: (
(_ldap_group_search_attribs = process.env.SHARELATEX_LDAP_GROUP_SEARCH_ATTRIBUTES) ?
(() => { try {
return JSON.parse(_ldap_group_search_attribs);
groupSearchAttributes: (_ldap_group_search_attribs =
process.env.SHARELATEX_LDAP_GROUP_SEARCH_ATTRIBUTES)
? (() => {
try {
return JSON.parse(_ldap_group_search_attribs)
} catch (error4) {
e = error4;
return console.error("could not parse SHARELATEX_LDAP_GROUP_SEARCH_ATTRIBUTES");
} })()
:
undefined
),
cache: process.env.SHARELATEX_LDAP_CACHE === 'true',
timeout: (
(_ldap_timeout = process.env.SHARELATEX_LDAP_TIMEOUT) ?
(() => { try {
return parseIntOrFail(_ldap_timeout);
} catch (error5) {
e = error5;
return console.error("Cannot parse SHARELATEX_LDAP_TIMEOUT");
} })()
:
undefined
),
connectTimeout: (
(_ldap_connect_timeout = process.env.SHARELATEX_LDAP_CONNECT_TIMEOUT) ?
(() => { try {
return parseIntOrFail(_ldap_connect_timeout);
} catch (error6) {
e = error6;
return console.error("Cannot parse SHARELATEX_LDAP_CONNECT_TIMEOUT");
} })()
:
undefined
e = error4
return console.error(
'could not parse SHARELATEX_LDAP_GROUP_SEARCH_ATTRIBUTES'
)
}
};
if (process.env.SHARELATEX_LDAP_TLS_OPTS_CA_PATH) {
let ca, ca_paths;
})()
: undefined,
cache: process.env.SHARELATEX_LDAP_CACHE === 'true',
timeout: (_ldap_timeout = process.env.SHARELATEX_LDAP_TIMEOUT)
? (() => {
try {
ca = JSON.parse(process.env.SHARELATEX_LDAP_TLS_OPTS_CA_PATH);
} catch (error7) {
e = error7;
console.error("could not parse SHARELATEX_LDAP_TLS_OPTS_CA_PATH, invalid JSON");
return parseIntOrFail(_ldap_timeout)
} catch (error5) {
e = error5
return console.error('Cannot parse SHARELATEX_LDAP_TIMEOUT')
}
})()
: undefined,
connectTimeout: (_ldap_connect_timeout =
process.env.SHARELATEX_LDAP_CONNECT_TIMEOUT)
? (() => {
try {
return parseIntOrFail(_ldap_connect_timeout)
} catch (error6) {
e = error6
return console.error(
'Cannot parse SHARELATEX_LDAP_CONNECT_TIMEOUT'
)
}
})()
: undefined,
},
}
if (typeof(ca) === 'string') {
ca_paths = [ca];
} else if ((typeof(ca) === 'object') && ((ca != null ? ca.length : undefined) != null)) {
ca_paths = ca;
if (process.env.SHARELATEX_LDAP_TLS_OPTS_CA_PATH) {
let ca, ca_paths
try {
ca = JSON.parse(process.env.SHARELATEX_LDAP_TLS_OPTS_CA_PATH)
} catch (error7) {
e = error7
console.error(
'could not parse SHARELATEX_LDAP_TLS_OPTS_CA_PATH, invalid JSON'
)
}
if (typeof ca === 'string') {
ca_paths = [ca]
} else if (
typeof ca === 'object' &&
(ca != null ? ca.length : undefined) != null
) {
ca_paths = ca
} else {
console.error("problem parsing SHARELATEX_LDAP_TLS_OPTS_CA_PATH");
console.error('problem parsing SHARELATEX_LDAP_TLS_OPTS_CA_PATH')
}
settings.ldap.server.tlsOptions = {
rejectUnauthorized: process.env.SHARELATEX_LDAP_TLS_OPTS_REJECT_UNAUTH === "true",
ca:ca_paths // e.g.'/etc/ldap/ca_certs.pem'
};
rejectUnauthorized:
process.env.SHARELATEX_LDAP_TLS_OPTS_REJECT_UNAUTH === 'true',
ca: ca_paths, // e.g.'/etc/ldap/ca_certs.pem'
}
}
}
if (process.env.SHARELATEX_SAML_ENTRYPOINT) {
// NOTE: see https://github.com/node-saml/passport-saml/blob/master/README.md for docs of `server` options
let _saml_additionalAuthorizeParams, _saml_additionalLogoutParams, _saml_additionalParams, _saml_expiration, _saml_skew;
settings.externalAuth = true;
let _saml_additionalAuthorizeParams,
_saml_additionalLogoutParams,
_saml_additionalParams,
_saml_expiration,
_saml_skew
settings.externalAuth = true
settings.saml = {
updateUserDetailsOnLogin: process.env.SHARELATEX_SAML_UPDATE_USER_DETAILS_ON_LOGIN === 'true',
updateUserDetailsOnLogin:
process.env.SHARELATEX_SAML_UPDATE_USER_DETAILS_ON_LOGIN === 'true',
identityServiceName: process.env.SHARELATEX_SAML_IDENTITY_SERVICE_NAME,
emailField: process.env.SHARELATEX_SAML_EMAIL_FIELD || process.env.SHARELATEX_SAML_EMAIL_FIELD_NAME,
emailField:
process.env.SHARELATEX_SAML_EMAIL_FIELD ||
process.env.SHARELATEX_SAML_EMAIL_FIELD_NAME,
firstNameField: process.env.SHARELATEX_SAML_FIRST_NAME_FIELD,
lastNameField: process.env.SHARELATEX_SAML_LAST_NAME_FIELD,
server: {
@ -531,162 +595,184 @@ if (process.env.SHARELATEX_SAML_ENTRYPOINT) {
decryptionCert: process.env.SHARELATEX_SAML_DECRYPTION_CERT,
signatureAlgorithm: process.env.SHARELATEX_SAML_SIGNATURE_ALGORITHM,
identifierFormat: process.env.SHARELATEX_SAML_IDENTIFIER_FORMAT,
attributeConsumingServiceIndex: process.env.SHARELATEX_SAML_ATTRIBUTE_CONSUMING_SERVICE_INDEX,
attributeConsumingServiceIndex:
process.env.SHARELATEX_SAML_ATTRIBUTE_CONSUMING_SERVICE_INDEX,
authnContext: process.env.SHARELATEX_SAML_AUTHN_CONTEXT,
authnRequestBinding: process.env.SHARELATEX_SAML_AUTHN_REQUEST_BINDING,
validateInResponseTo: process.env.SHARELATEX_SAML_VALIDATE_IN_RESPONSE_TO,
cacheProvider: process.env.SHARELATEX_SAML_CACHE_PROVIDER,
logoutUrl: process.env.SHARELATEX_SAML_LOGOUT_URL,
logoutCallbackUrl: process.env.SHARELATEX_SAML_LOGOUT_CALLBACK_URL,
disableRequestedAuthnContext: process.env.SHARELATEX_SAML_DISABLE_REQUESTED_AUTHN_CONTEXT === 'true',
disableRequestedAuthnContext:
process.env.SHARELATEX_SAML_DISABLE_REQUESTED_AUTHN_CONTEXT === 'true',
forceAuthn: process.env.SHARELATEX_SAML_FORCE_AUTHN === 'true',
skipRequestCompression: process.env.SHARELATEX_SAML_SKIP_REQUEST_COMPRESSION === 'true',
acceptedClockSkewMs: (
(_saml_skew = process.env.SHARELATEX_SAML_ACCEPTED_CLOCK_SKEW_MS) ?
(() => { try {
return parseIntOrFail(_saml_skew);
skipRequestCompression:
process.env.SHARELATEX_SAML_SKIP_REQUEST_COMPRESSION === 'true',
acceptedClockSkewMs: (_saml_skew =
process.env.SHARELATEX_SAML_ACCEPTED_CLOCK_SKEW_MS)
? (() => {
try {
return parseIntOrFail(_saml_skew)
} catch (error8) {
e = error8;
return console.error("Cannot parse SHARELATEX_SAML_ACCEPTED_CLOCK_SKEW_MS");
} })()
:
undefined
),
requestIdExpirationPeriodMs: (
(_saml_expiration = process.env.SHARELATEX_SAML_REQUEST_ID_EXPIRATION_PERIOD_MS) ?
(() => { try {
return parseIntOrFail(_saml_expiration);
} catch (error9) {
e = error9;
return console.error("Cannot parse SHARELATEX_SAML_REQUEST_ID_EXPIRATION_PERIOD_MS");
} })()
:
undefined
),
additionalParams: (
(_saml_additionalParams = process.env.SHARELATEX_SAML_ADDITIONAL_PARAMS) ?
(() => { try {
return JSON.parse(_saml_additionalParams);
} catch (error10) {
e = error10;
return console.error("Cannot parse SHARELATEX_SAML_ADDITIONAL_PARAMS");
} })()
:
undefined
),
additionalAuthorizeParams: (
(_saml_additionalAuthorizeParams = process.env.SHARELATEX_SAML_ADDITIONAL_AUTHORIZE_PARAMS) ?
(() => { try {
return JSON.parse(_saml_additionalAuthorizeParams );
} catch (error11) {
e = error11;
return console.error("Cannot parse SHARELATEX_SAML_ADDITIONAL_AUTHORIZE_PARAMS");
} })()
:
undefined
),
additionalLogoutParams: (
(_saml_additionalLogoutParams = process.env.SHARELATEX_SAML_ADDITIONAL_LOGOUT_PARAMS) ?
(() => { try {
return JSON.parse(_saml_additionalLogoutParams );
} catch (error12) {
e = error12;
return console.error("Cannot parse SHARELATEX_SAML_ADDITIONAL_LOGOUT_PARAMS");
} })()
:
undefined
e = error8
return console.error(
'Cannot parse SHARELATEX_SAML_ACCEPTED_CLOCK_SKEW_MS'
)
}
};
})()
: undefined,
requestIdExpirationPeriodMs: (_saml_expiration =
process.env.SHARELATEX_SAML_REQUEST_ID_EXPIRATION_PERIOD_MS)
? (() => {
try {
return parseIntOrFail(_saml_expiration)
} catch (error9) {
e = error9
return console.error(
'Cannot parse SHARELATEX_SAML_REQUEST_ID_EXPIRATION_PERIOD_MS'
)
}
})()
: undefined,
additionalParams: (_saml_additionalParams =
process.env.SHARELATEX_SAML_ADDITIONAL_PARAMS)
? (() => {
try {
return JSON.parse(_saml_additionalParams)
} catch (error10) {
e = error10
return console.error(
'Cannot parse SHARELATEX_SAML_ADDITIONAL_PARAMS'
)
}
})()
: undefined,
additionalAuthorizeParams: (_saml_additionalAuthorizeParams =
process.env.SHARELATEX_SAML_ADDITIONAL_AUTHORIZE_PARAMS)
? (() => {
try {
return JSON.parse(_saml_additionalAuthorizeParams)
} catch (error11) {
e = error11
return console.error(
'Cannot parse SHARELATEX_SAML_ADDITIONAL_AUTHORIZE_PARAMS'
)
}
})()
: undefined,
additionalLogoutParams: (_saml_additionalLogoutParams =
process.env.SHARELATEX_SAML_ADDITIONAL_LOGOUT_PARAMS)
? (() => {
try {
return JSON.parse(_saml_additionalLogoutParams)
} catch (error12) {
e = error12
return console.error(
'Cannot parse SHARELATEX_SAML_ADDITIONAL_LOGOUT_PARAMS'
)
}
})()
: undefined,
},
}
// SHARELATEX_SAML_CERT cannot be empty
// https://github.com/node-saml/passport-saml/commit/f6b1c885c0717f1083c664345556b535f217c102
if (process.env.SHARELATEX_SAML_CERT) {
settings.saml.server.cert = process.env.SHARELATEX_SAML_CERT;
settings.saml.server.privateCert = process.env.SHARELATEX_SAML_PRIVATE_CERT;
settings.saml.server.cert = process.env.SHARELATEX_SAML_CERT
settings.saml.server.privateCert = process.env.SHARELATEX_SAML_PRIVATE_CERT
}
}
// Compiler
// --------
if (process.env.SANDBOXED_COMPILES === "true") {
if (process.env.SANDBOXED_COMPILES === 'true') {
settings.clsi = {
dockerRunner: true,
docker: {
image: process.env.TEX_LIVE_DOCKER_IMAGE,
env: {
HOME: "/tmp",
PATH: process.env.COMPILER_PATH || "/usr/local/texlive/2015/bin/x86_64-linux:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
HOME: '/tmp',
PATH:
process.env.COMPILER_PATH ||
'/usr/local/texlive/2015/bin/x86_64-linux:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin',
},
user: 'www-data',
},
user: "www-data"
}
};
if ((settings.path == null)) {
settings.path = {};
if (settings.path == null) {
settings.path = {}
}
settings.path.synctexBaseDir = () => "/compile";
settings.path.synctexBaseDir = () => '/compile'
if (process.env.SANDBOXED_COMPILES_SIBLING_CONTAINERS === 'true') {
console.log("Using sibling containers for sandboxed compiles");
console.log('Using sibling containers for sandboxed compiles')
if (process.env.SANDBOXED_COMPILES_HOST_DIR) {
settings.path.sandboxedCompilesHostDir = process.env.SANDBOXED_COMPILES_HOST_DIR;
settings.path.sandboxedCompilesHostDir =
process.env.SANDBOXED_COMPILES_HOST_DIR
} else {
console.error('Sibling containers, but SANDBOXED_COMPILES_HOST_DIR not set');
console.error(
'Sibling containers, but SANDBOXED_COMPILES_HOST_DIR not set'
)
}
}
}
// Templates
// ---------
if (process.env.SHARELATEX_TEMPLATES_USER_ID) {
settings.templates = {
mountPointUrl: "/templates",
user_id: process.env.SHARELATEX_TEMPLATES_USER_ID
};
settings.templateLinks = parse(process.env.SHARELATEX_NEW_PROJECT_TEMPLATE_LINKS);
mountPointUrl: '/templates',
user_id: process.env.SHARELATEX_TEMPLATES_USER_ID,
}
settings.templateLinks = parse(
process.env.SHARELATEX_NEW_PROJECT_TEMPLATE_LINKS
)
}
// /Learn
// -------
if (process.env.SHARELATEX_PROXY_LEARN != null) {
settings.proxyLearn = parse(process.env.SHARELATEX_PROXY_LEARN);
settings.proxyLearn = parse(process.env.SHARELATEX_PROXY_LEARN)
}
// /References
// -----------
if (process.env.SHARELATEX_ELASTICSEARCH_URL != null) {
settings.references.elasticsearch =
{host: process.env.SHARELATEX_ELASTICSEARCH_URL};
settings.references.elasticsearch = {
host: process.env.SHARELATEX_ELASTICSEARCH_URL,
}
}
// TeX Live Images
// -----------
if (process.env.ALL_TEX_LIVE_DOCKER_IMAGES != null) {
allTexLiveDockerImages = process.env.ALL_TEX_LIVE_DOCKER_IMAGES.split(',');
allTexLiveDockerImages = process.env.ALL_TEX_LIVE_DOCKER_IMAGES.split(',')
}
if (process.env.ALL_TEX_LIVE_DOCKER_IMAGE_NAMES != null) {
allTexLiveDockerImageNames = process.env.ALL_TEX_LIVE_DOCKER_IMAGE_NAMES.split(',');
allTexLiveDockerImageNames =
process.env.ALL_TEX_LIVE_DOCKER_IMAGE_NAMES.split(',')
}
if (allTexLiveDockerImages != null) {
settings.allowedImageNames = [];
settings.allowedImageNames = []
for (let index = 0; index < allTexLiveDockerImages.length; index++) {
const fullImageName = allTexLiveDockerImages[index];
const imageName = Path.basename(fullImageName);
const imageDesc = (allTexLiveDockerImageNames != null) ? allTexLiveDockerImageNames[index] : imageName;
settings.allowedImageNames.push({ imageName, imageDesc });
const fullImageName = allTexLiveDockerImages[index]
const imageName = Path.basename(fullImageName)
const imageDesc =
allTexLiveDockerImageNames != null
? allTexLiveDockerImageNames[index]
: imageName
settings.allowedImageNames.push({ imageName, imageDesc })
}
}
// With lots of incoming and outgoing HTTP connections to different services,
// sometimes long running, it is a good idea to increase the default number
// of sockets that Node will hold open.
const http = require('http');
http.globalAgent.maxSockets = 300;
const https = require('https');
https.globalAgent.maxSockets = 300;
const http = require('http')
http.globalAgent.maxSockets = 300
const https = require('https')
https.globalAgent.maxSockets = 300
module.exports = settings;
module.exports = settings

View file

@ -12,34 +12,54 @@
*/
module.exports = function (grunt) {
grunt.registerTask('user:create-admin', "Create a user with the given email address and make them an admin. Update in place if the user already exists. Usage: grunt user:create-admin --email joe@example.com", function() {
const done = this.async();
const email = grunt.option("email");
if ((email == null)) {
console.error("Usage: grunt user:create-admin --email=joe@example.com");
process.exit(1);
grunt.registerTask(
'user:create-admin',
'Create a user with the given email address and make them an admin. Update in place if the user already exists. Usage: grunt user:create-admin --email joe@example.com',
function () {
const done = this.async()
const email = grunt.option('email')
if (email == null) {
console.error('Usage: grunt user:create-admin --email=joe@example.com')
process.exit(1)
}
const settings = require("settings-sharelatex");
const mongodb = require("../web/app/src/infrastructure/mongodb");
const UserRegistrationHandler = require("../web/app/src/Features/User/UserRegistrationHandler");
const OneTimeTokenHandler = require("../web/app/src/Features/Security/OneTimeTokenHandler");
return mongodb.waitForDb().then(() => UserRegistrationHandler.registerNewUser({
const settings = require('settings-sharelatex')
const mongodb = require('../web/app/src/infrastructure/mongodb')
const UserRegistrationHandler = require('../web/app/src/Features/User/UserRegistrationHandler')
const OneTimeTokenHandler = require('../web/app/src/Features/Security/OneTimeTokenHandler')
return mongodb.waitForDb().then(() =>
UserRegistrationHandler.registerNewUser(
{
email,
password: require("crypto").randomBytes(32).toString("hex")
}, function(error, user) {
if ((error != null) && ((error != null ? error.message : undefined) !== "EmailAlreadyRegistered")) {
throw error;
password: require('crypto').randomBytes(32).toString('hex'),
},
function (error, user) {
if (
error != null &&
(error != null ? error.message : undefined) !==
'EmailAlreadyRegistered'
) {
throw error
}
user.isAdmin = true;
user.isAdmin = true
return user.save(function (error) {
if (error != null) { throw error; }
const ONE_WEEK = 7 * 24 * 60 * 60; // seconds
return OneTimeTokenHandler.getNewToken("password", { expiresIn: ONE_WEEK, email:user.email, user_id: user._id.toString() }, function(err, token){
if (err != null) { return next(err); }
if (error != null) {
throw error
}
const ONE_WEEK = 7 * 24 * 60 * 60 // seconds
return OneTimeTokenHandler.getNewToken(
'password',
{
expiresIn: ONE_WEEK,
email: user.email,
user_id: user._id.toString(),
},
function (err, token) {
if (err != null) {
return next(err)
}
console.log("");
console.log('')
console.log(`\
Successfully created ${email} as an admin user.
@ -47,39 +67,50 @@ Please visit the following URL to set a password for ${email} and log in:
${settings.siteUrl}/user/password/set?passwordResetToken=${token}
\
`
);
return done();
});
});
}));
});
`)
return done()
}
)
})
}
)
)
}
)
return grunt.registerTask('user:delete', "deletes a user and all their data, Usage: grunt user:delete --email joe@example.com", function() {
const done = this.async();
const email = grunt.option("email");
if ((email == null)) {
console.error("Usage: grunt user:delete --email=joe@example.com");
process.exit(1);
return grunt.registerTask(
'user:delete',
'deletes a user and all their data, Usage: grunt user:delete --email joe@example.com',
function () {
const done = this.async()
const email = grunt.option('email')
if (email == null) {
console.error('Usage: grunt user:delete --email=joe@example.com')
process.exit(1)
}
const settings = require("settings-sharelatex");
const mongodb = require("../web/app/src/infrastructure/mongodb");
const UserGetter = require("../web/app/src/Features/User/UserGetter");
const UserDeleter = require("../web/app/src/Features/User/UserDeleter");
return mongodb.waitForDb().then(() => UserGetter.getUser({email}, function(error, user) {
const settings = require('settings-sharelatex')
const mongodb = require('../web/app/src/infrastructure/mongodb')
const UserGetter = require('../web/app/src/Features/User/UserGetter')
const UserDeleter = require('../web/app/src/Features/User/UserDeleter')
return mongodb.waitForDb().then(() =>
UserGetter.getUser({ email }, function (error, user) {
if (error != null) {
throw error;
throw error
}
if ((user == null)) {
console.log(`user ${email} not in database, potentially already deleted`);
return done();
if (user == null) {
console.log(
`user ${email} not in database, potentially already deleted`
)
return done()
}
return UserDeleter.deleteUser(user._id, function (err) {
if (err != null) {
throw err;
throw err
}
return done()
})
})
)
}
)
}
return done();
});
}));
});
};