mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-21 20:47:08 -05:00
[misc] run format:fix
This commit is contained in:
parent
d812c86c51
commit
d7c641eaf7
3 changed files with 993 additions and 839 deletions
|
@ -16,42 +16,36 @@
|
|||
* DS207: Consider shorter variations of null checks
|
||||
* Full docs: https://github.com/decaffeinate/decaffeinate/blob/master/docs/suggestions.md
|
||||
*/
|
||||
const coffee = require("coffee-script");
|
||||
const fs = require("fs");
|
||||
const {
|
||||
spawn
|
||||
} = require("child_process");
|
||||
const {
|
||||
exec
|
||||
} = require("child_process");
|
||||
const rimraf = require("rimraf");
|
||||
const Path = require("path");
|
||||
const semver = require("semver");
|
||||
const knox = require("knox");
|
||||
const crypto = require("crypto");
|
||||
const async = require("async");
|
||||
const settings = require("settings-sharelatex");
|
||||
const _ = require("underscore");
|
||||
const coffee = require('coffee-script')
|
||||
const fs = require('fs')
|
||||
const { spawn } = require('child_process')
|
||||
const { exec } = require('child_process')
|
||||
const rimraf = require('rimraf')
|
||||
const Path = require('path')
|
||||
const semver = require('semver')
|
||||
const knox = require('knox')
|
||||
const crypto = require('crypto')
|
||||
const async = require('async')
|
||||
const settings = require('settings-sharelatex')
|
||||
const _ = require('underscore')
|
||||
|
||||
|
||||
const SERVICES = require("./config/services");
|
||||
const SERVICES = require('./config/services')
|
||||
|
||||
module.exports = function (grunt) {
|
||||
let Helpers;
|
||||
let service;
|
||||
grunt.loadNpmTasks('grunt-bunyan');
|
||||
grunt.loadNpmTasks('grunt-execute');
|
||||
grunt.loadNpmTasks('grunt-available-tasks');
|
||||
grunt.loadNpmTasks('grunt-concurrent');
|
||||
grunt.loadNpmTasks("grunt-contrib-coffee");
|
||||
grunt.loadNpmTasks("grunt-shell");
|
||||
let Helpers
|
||||
let service
|
||||
grunt.loadNpmTasks('grunt-bunyan')
|
||||
grunt.loadNpmTasks('grunt-execute')
|
||||
grunt.loadNpmTasks('grunt-available-tasks')
|
||||
grunt.loadNpmTasks('grunt-concurrent')
|
||||
grunt.loadNpmTasks('grunt-contrib-coffee')
|
||||
grunt.loadNpmTasks('grunt-shell')
|
||||
|
||||
grunt.task.loadTasks("./tasks");
|
||||
grunt.task.loadTasks('./tasks')
|
||||
|
||||
const execute = {};
|
||||
const execute = {}
|
||||
for (service of Array.from(SERVICES)) {
|
||||
execute[service.name] =
|
||||
{src: `${service.name}/app.js`};
|
||||
execute[service.name] = { src: `${service.name}/app.js` }
|
||||
}
|
||||
|
||||
grunt.initConfig({
|
||||
|
@ -59,178 +53,221 @@ module.exports = function(grunt) {
|
|||
|
||||
concurrent: {
|
||||
all: {
|
||||
tasks: (((() => {
|
||||
const result = [];
|
||||
for (service of Array.from(SERVICES)) { result.push(`run:${service.name}`);
|
||||
tasks: (() => {
|
||||
const result = []
|
||||
for (service of Array.from(SERVICES)) {
|
||||
result.push(`run:${service.name}`)
|
||||
}
|
||||
return result;
|
||||
})())),
|
||||
return result
|
||||
})(),
|
||||
options: {
|
||||
limit: SERVICES.length,
|
||||
logConcurrentOutput: true
|
||||
logConcurrentOutput: true,
|
||||
},
|
||||
},
|
||||
},
|
||||
|
||||
availabletasks: {
|
||||
tasks: {
|
||||
options: {
|
||||
filter: 'exclude',
|
||||
tasks: [
|
||||
'concurrent',
|
||||
'execute',
|
||||
'bunyan',
|
||||
'availabletasks'
|
||||
],
|
||||
tasks: ['concurrent', 'execute', 'bunyan', 'availabletasks'],
|
||||
groups: {
|
||||
"Run tasks": [
|
||||
"run",
|
||||
"run:all",
|
||||
"default"
|
||||
].concat(((() => {
|
||||
const result1 = [];
|
||||
for (service of Array.from(SERVICES)) { result1.push(`run:${service.name}`);
|
||||
'Run tasks': ['run', 'run:all', 'default'].concat(
|
||||
(() => {
|
||||
const result1 = []
|
||||
for (service of Array.from(SERVICES)) {
|
||||
result1.push(`run:${service.name}`)
|
||||
}
|
||||
return result1;
|
||||
})())),
|
||||
"Misc": [
|
||||
"help"
|
||||
return result1
|
||||
})()
|
||||
),
|
||||
Misc: ['help'],
|
||||
'Install tasks': (() => {
|
||||
const result2 = []
|
||||
for (service of Array.from(SERVICES)) {
|
||||
result2.push(`install:${service.name}`)
|
||||
}
|
||||
return result2
|
||||
})().concat(['install:all', 'install']),
|
||||
'Update tasks': (() => {
|
||||
const result3 = []
|
||||
for (service of Array.from(SERVICES)) {
|
||||
result3.push(`update:${service.name}`)
|
||||
}
|
||||
return result3
|
||||
})().concat(['update:all', 'update']),
|
||||
Checks: [
|
||||
'check',
|
||||
'check:redis',
|
||||
'check:latexmk',
|
||||
'check:s3',
|
||||
'check:make',
|
||||
'check:mongo',
|
||||
],
|
||||
"Install tasks": ((() => {
|
||||
const result2 = [];
|
||||
for (service of Array.from(SERVICES)) { result2.push(`install:${service.name}`);
|
||||
}
|
||||
return result2;
|
||||
})()).concat(["install:all", "install"]),
|
||||
"Update tasks": ((() => {
|
||||
const result3 = [];
|
||||
for (service of Array.from(SERVICES)) { result3.push(`update:${service.name}`);
|
||||
}
|
||||
return result3;
|
||||
})()).concat(["update:all", "update"]),
|
||||
"Checks": ["check", "check:redis", "check:latexmk", "check:s3", "check:make", "check:mongo"]
|
||||
}
|
||||
}
|
||||
}
|
||||
}});
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
})
|
||||
|
||||
for (service of Array.from(SERVICES)) {
|
||||
((service => grunt.registerTask(`install:${service.name}`, `Download and set up the ${service.name} service`, function() {
|
||||
const done = this.async();
|
||||
return Helpers.installService(service, done);
|
||||
})))(service);
|
||||
;(service =>
|
||||
grunt.registerTask(
|
||||
`install:${service.name}`,
|
||||
`Download and set up the ${service.name} service`,
|
||||
function () {
|
||||
const done = this.async()
|
||||
return Helpers.installService(service, done)
|
||||
}
|
||||
))(service)
|
||||
}
|
||||
|
||||
|
||||
|
||||
grunt.registerTask('install:all', "Download and set up all ShareLaTeX services",
|
||||
[].concat(
|
||||
((() => {
|
||||
const result4 = [];
|
||||
for (service of Array.from(SERVICES)) { result4.push(`install:${service.name}`);
|
||||
grunt.registerTask(
|
||||
'install:all',
|
||||
'Download and set up all ShareLaTeX services',
|
||||
[]
|
||||
.concat(
|
||||
(() => {
|
||||
const result4 = []
|
||||
for (service of Array.from(SERVICES)) {
|
||||
result4.push(`install:${service.name}`)
|
||||
}
|
||||
return result4;
|
||||
})())
|
||||
).concat(['postinstall'])
|
||||
);
|
||||
|
||||
grunt.registerTask('install', 'install:all');
|
||||
grunt.registerTask('postinstall', 'Explain postinstall steps', function() {
|
||||
return Helpers.postinstallMessage(this.async());
|
||||
});
|
||||
|
||||
grunt.registerTask('update:all', "Checkout and update all ShareLaTeX services",
|
||||
["check:make"].concat(
|
||||
((() => {
|
||||
const result5 = [];
|
||||
for (service of Array.from(SERVICES)) { result5.push(`update:${service.name}`);
|
||||
}
|
||||
return result5;
|
||||
})())
|
||||
return result4
|
||||
})()
|
||||
)
|
||||
.concat(['postinstall'])
|
||||
)
|
||||
);
|
||||
grunt.registerTask('update', 'update:all');
|
||||
grunt.registerTask('run', "Run all of the sharelatex processes", ['concurrent:all']);
|
||||
grunt.registerTask('run:all', 'run');
|
||||
|
||||
grunt.registerTask('help', 'Display this help list', 'availabletasks');
|
||||
grunt.registerTask('default', 'run');
|
||||
grunt.registerTask('install', 'install:all')
|
||||
grunt.registerTask('postinstall', 'Explain postinstall steps', function () {
|
||||
return Helpers.postinstallMessage(this.async())
|
||||
})
|
||||
|
||||
grunt.registerTask("check:redis", "Check that redis is installed and running", function() {
|
||||
return Helpers.checkRedisConnect(this.async());
|
||||
});
|
||||
grunt.registerTask(
|
||||
'update:all',
|
||||
'Checkout and update all ShareLaTeX services',
|
||||
['check:make'].concat(
|
||||
(() => {
|
||||
const result5 = []
|
||||
for (service of Array.from(SERVICES)) {
|
||||
result5.push(`update:${service.name}`)
|
||||
}
|
||||
return result5
|
||||
})()
|
||||
)
|
||||
)
|
||||
grunt.registerTask('update', 'update:all')
|
||||
grunt.registerTask('run', 'Run all of the sharelatex processes', [
|
||||
'concurrent:all',
|
||||
])
|
||||
grunt.registerTask('run:all', 'run')
|
||||
|
||||
grunt.registerTask("check:mongo", "Check that mongo is installed", function() {
|
||||
return Helpers.checkMongoConnect(this.async());
|
||||
});
|
||||
grunt.registerTask('help', 'Display this help list', 'availabletasks')
|
||||
grunt.registerTask('default', 'run')
|
||||
|
||||
grunt.registerTask("check", "Check that you have the required dependencies installed", ["check:redis", "check:mongo", "check:make"]);
|
||||
grunt.registerTask(
|
||||
'check:redis',
|
||||
'Check that redis is installed and running',
|
||||
function () {
|
||||
return Helpers.checkRedisConnect(this.async())
|
||||
}
|
||||
)
|
||||
|
||||
grunt.registerTask("check:make", "Check that make is installed", function() {
|
||||
return Helpers.checkMake(this.async());
|
||||
});
|
||||
grunt.registerTask(
|
||||
'check:mongo',
|
||||
'Check that mongo is installed',
|
||||
function () {
|
||||
return Helpers.checkMongoConnect(this.async())
|
||||
}
|
||||
)
|
||||
|
||||
grunt.registerTask(
|
||||
'check',
|
||||
'Check that you have the required dependencies installed',
|
||||
['check:redis', 'check:mongo', 'check:make']
|
||||
)
|
||||
|
||||
return Helpers = {
|
||||
grunt.registerTask('check:make', 'Check that make is installed', function () {
|
||||
return Helpers.checkMake(this.async())
|
||||
})
|
||||
|
||||
return (Helpers = {
|
||||
installService(service, callback) {
|
||||
if (callback == null) { callback = function(error) {}; }
|
||||
console.log(`Installing ${service.name}`);
|
||||
if (callback == null) {
|
||||
callback = function (error) {}
|
||||
}
|
||||
console.log(`Installing ${service.name}`)
|
||||
return Helpers.cloneGitRepo(service, function (error) {
|
||||
if (error != null) {
|
||||
return callback(error);
|
||||
return callback(error)
|
||||
} else {
|
||||
return callback();
|
||||
return callback()
|
||||
}
|
||||
});
|
||||
})
|
||||
},
|
||||
|
||||
cloneGitRepo(service, callback) {
|
||||
if (callback == null) { callback = function(error) {}; }
|
||||
const repo_src = service.repo;
|
||||
const dir = service.name;
|
||||
if (callback == null) {
|
||||
callback = function (error) {}
|
||||
}
|
||||
const repo_src = service.repo
|
||||
const dir = service.name
|
||||
if (!fs.existsSync(dir)) {
|
||||
const proc = spawn("git", [
|
||||
"clone",
|
||||
repo_src,
|
||||
dir
|
||||
], {stdio: "inherit"});
|
||||
return proc.on("close", () => Helpers.checkoutVersion(service, callback));
|
||||
const proc = spawn('git', ['clone', repo_src, dir], {
|
||||
stdio: 'inherit',
|
||||
})
|
||||
return proc.on('close', () =>
|
||||
Helpers.checkoutVersion(service, callback)
|
||||
)
|
||||
} else {
|
||||
console.log(`${dir} already installed, skipping.`);
|
||||
return callback();
|
||||
console.log(`${dir} already installed, skipping.`)
|
||||
return callback()
|
||||
}
|
||||
},
|
||||
|
||||
checkoutVersion(service, callback) {
|
||||
if (callback == null) { callback = function(error) {}; }
|
||||
const dir = service.name;
|
||||
grunt.log.write(`checking out ${service.name} ${service.version}`);
|
||||
const proc = spawn("git", ["checkout", service.version], {stdio: "inherit", cwd: dir});
|
||||
return proc.on("close", () => callback());
|
||||
if (callback == null) {
|
||||
callback = function (error) {}
|
||||
}
|
||||
const dir = service.name
|
||||
grunt.log.write(`checking out ${service.name} ${service.version}`)
|
||||
const proc = spawn('git', ['checkout', service.version], {
|
||||
stdio: 'inherit',
|
||||
cwd: dir,
|
||||
})
|
||||
return proc.on('close', () => callback())
|
||||
},
|
||||
|
||||
postinstallMessage(callback) {
|
||||
if (callback == null) { callback = function(error) {}; }
|
||||
if (callback == null) {
|
||||
callback = function (error) {}
|
||||
}
|
||||
grunt.log.write(`\
|
||||
Services cloned:
|
||||
${(() => {
|
||||
const result6 = [];
|
||||
for (service of Array.from(SERVICES)) { result6.push(service.name);
|
||||
const result6 = []
|
||||
for (service of Array.from(SERVICES)) {
|
||||
result6.push(service.name)
|
||||
}
|
||||
return result6;
|
||||
return result6
|
||||
})()}
|
||||
To install services run:
|
||||
$ source bin/install-services
|
||||
This will install the required node versions and run \`npm install\` for each service.
|
||||
See https://github.com/sharelatex/sharelatex/pull/549 for more info.\
|
||||
`
|
||||
);
|
||||
return callback();
|
||||
`)
|
||||
return callback()
|
||||
},
|
||||
|
||||
checkMake(callback) {
|
||||
if (callback == null) { callback = function(error) {}; }
|
||||
grunt.log.write("Checking make is installed... ");
|
||||
return exec("make --version", function(error, stdout, stderr) {
|
||||
if ((error != null) && error.message.match("not found")) {
|
||||
grunt.log.error("FAIL.");
|
||||
if (callback == null) {
|
||||
callback = function (error) {}
|
||||
}
|
||||
grunt.log.write('Checking make is installed... ')
|
||||
return exec('make --version', function (error, stdout, stderr) {
|
||||
if (error != null && error.message.match('not found')) {
|
||||
grunt.log.error('FAIL.')
|
||||
grunt.log.errorlns(`\
|
||||
Either make is not installed or is not in your path.
|
||||
|
||||
|
@ -238,31 +275,32 @@ On Ubuntu you can install make with:
|
|||
|
||||
sudo apt-get install build-essential
|
||||
\
|
||||
`
|
||||
);
|
||||
return callback(error);
|
||||
`)
|
||||
return callback(error)
|
||||
} else if (error != null) {
|
||||
return callback(error);
|
||||
return callback(error)
|
||||
} else {
|
||||
grunt.log.write("OK.");
|
||||
return callback();
|
||||
grunt.log.write('OK.')
|
||||
return callback()
|
||||
}
|
||||
});
|
||||
})
|
||||
},
|
||||
checkMongoConnect(callback) {
|
||||
if (callback == null) { callback = function(error) {}; }
|
||||
grunt.log.write("Checking can connect to mongo");
|
||||
const mongojs = require("mongojs");
|
||||
const db = mongojs(settings.mongo.url, ["tags"]);
|
||||
if (callback == null) {
|
||||
callback = function (error) {}
|
||||
}
|
||||
grunt.log.write('Checking can connect to mongo')
|
||||
const mongojs = require('mongojs')
|
||||
const db = mongojs(settings.mongo.url, ['tags'])
|
||||
db.runCommand({ ping: 1 }, function (err, res) {
|
||||
if (!err && res.ok) {
|
||||
grunt.log.write("OK.");
|
||||
grunt.log.write('OK.')
|
||||
}
|
||||
return callback();
|
||||
});
|
||||
return callback()
|
||||
})
|
||||
return db.on('error', function (err) {
|
||||
err = "Can not connect to mongodb";
|
||||
grunt.log.error("FAIL.");
|
||||
err = 'Can not connect to mongodb'
|
||||
grunt.log.error('FAIL.')
|
||||
grunt.log.errorlns(`\
|
||||
!!!!!!!!!!!!!! MONGO ERROR !!!!!!!!!!!!!!
|
||||
|
||||
|
@ -271,29 +309,30 @@ ShareLaTeX can not talk to the mongodb instance
|
|||
Check the mongodb instance is running and accessible on env var SHARELATEX_MONGO_URL
|
||||
|
||||
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\
|
||||
`
|
||||
);
|
||||
throw new Error("Can not connect to Mongodb");
|
||||
return callback(err);
|
||||
});
|
||||
`)
|
||||
throw new Error('Can not connect to Mongodb')
|
||||
return callback(err)
|
||||
})
|
||||
},
|
||||
|
||||
checkRedisConnect(callback) {
|
||||
if (callback == null) { callback = function(error) {}; }
|
||||
grunt.log.write("Checking can connect to redis\n");
|
||||
const rclient = require("redis").createClient(settings.redis.web);
|
||||
if (callback == null) {
|
||||
callback = function (error) {}
|
||||
}
|
||||
grunt.log.write('Checking can connect to redis\n')
|
||||
const rclient = require('redis').createClient(settings.redis.web)
|
||||
|
||||
rclient.ping(function (err, res) {
|
||||
if ((err == null)) {
|
||||
grunt.log.write("OK.");
|
||||
if (err == null) {
|
||||
grunt.log.write('OK.')
|
||||
} else {
|
||||
throw new Error("Can not connect to redis");
|
||||
throw new Error('Can not connect to redis')
|
||||
}
|
||||
return callback();
|
||||
});
|
||||
return callback()
|
||||
})
|
||||
const errorHandler = _.once(function (err) {
|
||||
err = "Can not connect to redis";
|
||||
grunt.log.error("FAIL.");
|
||||
err = 'Can not connect to redis'
|
||||
grunt.log.error('FAIL.')
|
||||
grunt.log.errorlns(`\
|
||||
!!!!!!!!!!!!!! REDIS ERROR !!!!!!!!!!!!!!
|
||||
|
||||
|
@ -302,19 +341,17 @@ ShareLaTeX can not talk to the redis instance
|
|||
Check the redis instance is running and accessible on env var SHARELATEX_REDIS_HOST
|
||||
|
||||
!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!\
|
||||
`
|
||||
);
|
||||
throw new Error("Can not connect to redis");
|
||||
return callback(err);
|
||||
});
|
||||
return rclient.on('error', errorHandler);
|
||||
`)
|
||||
throw new Error('Can not connect to redis')
|
||||
return callback(err)
|
||||
})
|
||||
return rclient.on('error', errorHandler)
|
||||
},
|
||||
})
|
||||
}
|
||||
};
|
||||
};
|
||||
|
||||
|
||||
|
||||
|
||||
function __guard__(value, transform) {
|
||||
return (typeof value !== 'undefined' && value !== null) ? transform(value) : undefined;
|
||||
return typeof value !== 'undefined' && value !== null
|
||||
? transform(value)
|
||||
: undefined
|
||||
}
|
||||
|
|
|
@ -12,46 +12,45 @@
|
|||
* DS207: Consider shorter variations of null checks
|
||||
* Full docs: https://github.com/decaffeinate/decaffeinate/blob/master/docs/suggestions.md
|
||||
*/
|
||||
let allTexLiveDockerImageNames, allTexLiveDockerImages, redisConfig, siteUrl;
|
||||
let e;
|
||||
const Path = require('path');
|
||||
let allTexLiveDockerImageNames, allTexLiveDockerImages, redisConfig, siteUrl
|
||||
let e
|
||||
const Path = require('path')
|
||||
|
||||
// These credentials are used for authenticating api requests
|
||||
// between services that may need to go over public channels
|
||||
const httpAuthUser = "sharelatex";
|
||||
const httpAuthPass = process.env.WEB_API_PASSWORD;
|
||||
const httpAuthUsers = {};
|
||||
httpAuthUsers[httpAuthUser] = httpAuthPass;
|
||||
const httpAuthUser = 'sharelatex'
|
||||
const httpAuthPass = process.env.WEB_API_PASSWORD
|
||||
const httpAuthUsers = {}
|
||||
httpAuthUsers[httpAuthUser] = httpAuthPass
|
||||
|
||||
const parse = function (option) {
|
||||
if (option != null) {
|
||||
try {
|
||||
const opt = JSON.parse(option);
|
||||
return opt;
|
||||
const opt = JSON.parse(option)
|
||||
return opt
|
||||
} catch (err) {
|
||||
throw new Error(`problem parsing ${option}, invalid JSON`);
|
||||
throw new Error(`problem parsing ${option}, invalid JSON`)
|
||||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
const parseIntOrFail = function (value) {
|
||||
const parsedValue = parseInt(value, 10);
|
||||
const parsedValue = parseInt(value, 10)
|
||||
if (isNaN(parsedValue)) {
|
||||
throw new Error(`'${value}' is an invalid integer`);
|
||||
throw new Error(`'${value}' is an invalid integer`)
|
||||
}
|
||||
return parsedValue
|
||||
}
|
||||
return parsedValue;
|
||||
};
|
||||
|
||||
const DATA_DIR = '/var/lib/sharelatex/data';
|
||||
const TMP_DIR = '/var/lib/sharelatex/tmp';
|
||||
const DATA_DIR = '/var/lib/sharelatex/data'
|
||||
const TMP_DIR = '/var/lib/sharelatex/tmp'
|
||||
|
||||
const settings = {
|
||||
|
||||
clsi: {
|
||||
optimiseInDocker: process.env.OPTIMISE_PDF === 'true'
|
||||
optimiseInDocker: process.env.OPTIMISE_PDF === 'true',
|
||||
},
|
||||
|
||||
brandPrefix: "",
|
||||
brandPrefix: '',
|
||||
|
||||
allowAnonymousReadAndWriteSharing:
|
||||
process.env.SHARELATEX_ALLOW_ANONYMOUS_READ_AND_WRITE_SHARING === 'true',
|
||||
|
@ -66,7 +65,7 @@ const settings = {
|
|||
//
|
||||
// The following works out of the box with Mongo's default settings:
|
||||
mongo: {
|
||||
url : process.env.SHARELATEX_MONGO_URL || 'mongodb://dockerhost/sharelatex'
|
||||
url: process.env.SHARELATEX_MONGO_URL || 'mongodb://dockerhost/sharelatex',
|
||||
},
|
||||
|
||||
// Redis is used in ShareLaTeX for high volume queries, like real-time
|
||||
|
@ -75,36 +74,72 @@ const settings = {
|
|||
// The following config will work with Redis's default settings:
|
||||
redis: {
|
||||
web: (redisConfig = {
|
||||
host: process.env.SHARELATEX_REDIS_HOST || "dockerhost",
|
||||
port: process.env.SHARELATEX_REDIS_PORT || "6379",
|
||||
host: process.env.SHARELATEX_REDIS_HOST || 'dockerhost',
|
||||
port: process.env.SHARELATEX_REDIS_PORT || '6379',
|
||||
password: process.env.SHARELATEX_REDIS_PASS || undefined,
|
||||
key_schema: {
|
||||
// document-updater
|
||||
blockingKey({doc_id}) { return `Blocking:${doc_id}`; },
|
||||
docLines({doc_id}) { return `doclines:${doc_id}`; },
|
||||
docOps({doc_id}) { return `DocOps:${doc_id}`; },
|
||||
docVersion({doc_id}) { return `DocVersion:${doc_id}`; },
|
||||
docHash({doc_id}) { return `DocHash:${doc_id}`; },
|
||||
projectKey({doc_id}) { return `ProjectId:${doc_id}`; },
|
||||
docsInProject({project_id}) { return `DocsIn:${project_id}`; },
|
||||
ranges({doc_id}) { return `Ranges:${doc_id}`; },
|
||||
blockingKey({ doc_id }) {
|
||||
return `Blocking:${doc_id}`
|
||||
},
|
||||
docLines({ doc_id }) {
|
||||
return `doclines:${doc_id}`
|
||||
},
|
||||
docOps({ doc_id }) {
|
||||
return `DocOps:${doc_id}`
|
||||
},
|
||||
docVersion({ doc_id }) {
|
||||
return `DocVersion:${doc_id}`
|
||||
},
|
||||
docHash({ doc_id }) {
|
||||
return `DocHash:${doc_id}`
|
||||
},
|
||||
projectKey({ doc_id }) {
|
||||
return `ProjectId:${doc_id}`
|
||||
},
|
||||
docsInProject({ project_id }) {
|
||||
return `DocsIn:${project_id}`
|
||||
},
|
||||
ranges({ doc_id }) {
|
||||
return `Ranges:${doc_id}`
|
||||
},
|
||||
// document-updater:realtime
|
||||
pendingUpdates({doc_id}) { return `PendingUpdates:${doc_id}`; },
|
||||
pendingUpdates({ doc_id }) {
|
||||
return `PendingUpdates:${doc_id}`
|
||||
},
|
||||
// document-updater:history
|
||||
uncompressedHistoryOps({doc_id}) { return `UncompressedHistoryOps:${doc_id}`; },
|
||||
docsWithHistoryOps({project_id}) { return `DocsWithHistoryOps:${project_id}`; },
|
||||
uncompressedHistoryOps({ doc_id }) {
|
||||
return `UncompressedHistoryOps:${doc_id}`
|
||||
},
|
||||
docsWithHistoryOps({ project_id }) {
|
||||
return `DocsWithHistoryOps:${project_id}`
|
||||
},
|
||||
// document-updater:lock
|
||||
blockingKey({doc_id}) { return `Blocking:${doc_id}`; },
|
||||
blockingKey({ doc_id }) {
|
||||
return `Blocking:${doc_id}`
|
||||
},
|
||||
// track-changes:lock
|
||||
historyLock({doc_id}) { return `HistoryLock:${doc_id}`; },
|
||||
historyIndexLock({project_id}) { return `HistoryIndexLock:${project_id}`; },
|
||||
historyLock({ doc_id }) {
|
||||
return `HistoryLock:${doc_id}`
|
||||
},
|
||||
historyIndexLock({ project_id }) {
|
||||
return `HistoryIndexLock:${project_id}`
|
||||
},
|
||||
// track-changes:history
|
||||
uncompressedHistoryOps({doc_id}) { return `UncompressedHistoryOps:${doc_id}`; },
|
||||
docsWithHistoryOps({project_id}) { return `DocsWithHistoryOps:${project_id}`; },
|
||||
uncompressedHistoryOps({ doc_id }) {
|
||||
return `UncompressedHistoryOps:${doc_id}`
|
||||
},
|
||||
docsWithHistoryOps({ project_id }) {
|
||||
return `DocsWithHistoryOps:${project_id}`
|
||||
},
|
||||
// realtime
|
||||
clientsInProject({project_id}) { return `clients_in_project:${project_id}`; },
|
||||
connectedUser({project_id, client_id}){ return `connected_user:${project_id}:${client_id}`; }
|
||||
}
|
||||
clientsInProject({ project_id }) {
|
||||
return `clients_in_project:${project_id}`
|
||||
},
|
||||
connectedUser({ project_id, client_id }) {
|
||||
return `connected_user:${project_id}:${client_id}`
|
||||
},
|
||||
},
|
||||
}),
|
||||
fairy: redisConfig,
|
||||
// track-changes and document-updater
|
||||
|
@ -115,7 +150,7 @@ const settings = {
|
|||
websessions: redisConfig,
|
||||
api: redisConfig,
|
||||
pubsub: redisConfig,
|
||||
project_history: redisConfig
|
||||
project_history: redisConfig,
|
||||
},
|
||||
|
||||
// The compile server (the clsi) uses a SQL database to cache files and
|
||||
|
@ -129,12 +164,12 @@ const settings = {
|
|||
//
|
||||
mysql: {
|
||||
clsi: {
|
||||
database: "clsi",
|
||||
username: "clsi",
|
||||
password: "",
|
||||
dialect: "sqlite",
|
||||
storage: Path.join(DATA_DIR, "db.sqlite")
|
||||
}
|
||||
database: 'clsi',
|
||||
username: 'clsi',
|
||||
password: '',
|
||||
dialect: 'sqlite',
|
||||
storage: Path.join(DATA_DIR, 'db.sqlite'),
|
||||
},
|
||||
},
|
||||
|
||||
// File storage
|
||||
|
@ -143,11 +178,11 @@ const settings = {
|
|||
// ShareLaTeX can store binary files like images either locally or in Amazon
|
||||
// S3. The default is locally:
|
||||
filestore: {
|
||||
backend: "fs",
|
||||
backend: 'fs',
|
||||
stores: {
|
||||
user_files: Path.join(DATA_DIR, "user_files"),
|
||||
template_files: Path.join(DATA_DIR, "template_files")
|
||||
}
|
||||
user_files: Path.join(DATA_DIR, 'user_files'),
|
||||
template_files: Path.join(DATA_DIR, 'template_files'),
|
||||
},
|
||||
},
|
||||
|
||||
// To use Amazon S3 as a storage backend, comment out the above config, and
|
||||
|
@ -163,7 +198,7 @@ const settings = {
|
|||
//
|
||||
|
||||
trackchanges: {
|
||||
continueOnError: true
|
||||
continueOnError: true,
|
||||
},
|
||||
|
||||
// Local disk caching
|
||||
|
@ -172,15 +207,15 @@ const settings = {
|
|||
// If we ever need to write something to disk (e.g. incoming requests
|
||||
// that need processing but may be too big for memory), then write
|
||||
// them to disk here:
|
||||
dumpFolder: Path.join(TMP_DIR, "dumpFolder"),
|
||||
dumpFolder: Path.join(TMP_DIR, 'dumpFolder'),
|
||||
// Where to write uploads before they are processed
|
||||
uploadFolder: Path.join(TMP_DIR, "uploads"),
|
||||
uploadFolder: Path.join(TMP_DIR, 'uploads'),
|
||||
// Where to write the project to disk before running LaTeX on it
|
||||
compilesDir: Path.join(DATA_DIR, "compiles"),
|
||||
compilesDir: Path.join(DATA_DIR, 'compiles'),
|
||||
// Where to cache downloaded URLs for the CLSI
|
||||
clsiCacheDir: Path.join(DATA_DIR, "cache"),
|
||||
clsiCacheDir: Path.join(DATA_DIR, 'cache'),
|
||||
// Where to write the output files to disk after running LaTeX
|
||||
outputDir: Path.join(DATA_DIR, "output")
|
||||
outputDir: Path.join(DATA_DIR, 'output'),
|
||||
},
|
||||
|
||||
// Server Config
|
||||
|
@ -191,23 +226,27 @@ const settings = {
|
|||
siteUrl: (siteUrl = process.env.SHARELATEX_SITE_URL || 'http://localhost'),
|
||||
|
||||
// The name this is used to describe your ShareLaTeX Installation
|
||||
appName: process.env.SHARELATEX_APP_NAME || "ShareLaTeX (Community Edition)",
|
||||
appName: process.env.SHARELATEX_APP_NAME || 'ShareLaTeX (Community Edition)',
|
||||
|
||||
restrictInvitesToExistingAccounts: process.env.SHARELATEX_RESTRICT_INVITES_TO_EXISTING_ACCOUNTS === 'true',
|
||||
restrictInvitesToExistingAccounts:
|
||||
process.env.SHARELATEX_RESTRICT_INVITES_TO_EXISTING_ACCOUNTS === 'true',
|
||||
|
||||
nav: {
|
||||
title: process.env.SHARELATEX_NAV_TITLE || process.env.SHARELATEX_APP_NAME || "ShareLaTeX Community Edition"
|
||||
title:
|
||||
process.env.SHARELATEX_NAV_TITLE ||
|
||||
process.env.SHARELATEX_APP_NAME ||
|
||||
'ShareLaTeX Community Edition',
|
||||
},
|
||||
|
||||
|
||||
// The email address which users will be directed to as the main point of
|
||||
// contact for this installation of ShareLaTeX.
|
||||
adminEmail: process.env.SHARELATEX_ADMIN_EMAIL || "placeholder@example.com",
|
||||
adminEmail: process.env.SHARELATEX_ADMIN_EMAIL || 'placeholder@example.com',
|
||||
|
||||
// If provided, a sessionSecret is used to sign cookies so that they cannot be
|
||||
// spoofed. This is recommended.
|
||||
security: {
|
||||
sessionSecret: process.env.SHARELATEX_SESSION_SECRET || process.env.CRYPTO_RANDOM
|
||||
sessionSecret:
|
||||
process.env.SHARELATEX_SESSION_SECRET || process.env.CRYPTO_RANDOM,
|
||||
},
|
||||
|
||||
// These credentials are used for authenticating api requests
|
||||
|
@ -224,7 +263,7 @@ const settings = {
|
|||
|
||||
// If you are running ShareLaTeX over https, set this to true to send the
|
||||
// cookie with a secure flag (recommended).
|
||||
secureCookie: (process.env.SHARELATEX_SECURE_COOKIE != null),
|
||||
secureCookie: process.env.SHARELATEX_SECURE_COOKIE != null,
|
||||
|
||||
// If you are running ShareLaTeX behind a proxy (like Apache, Nginx, etc)
|
||||
// then set this to true to allow it to correctly detect the forwarded IP
|
||||
|
@ -234,22 +273,25 @@ const settings = {
|
|||
|
||||
i18n: {
|
||||
subdomainLang: {
|
||||
www: {lngCode:process.env.SHARELATEX_SITE_LANGUAGE || "en", url: siteUrl}
|
||||
www: {
|
||||
lngCode: process.env.SHARELATEX_SITE_LANGUAGE || 'en',
|
||||
url: siteUrl,
|
||||
},
|
||||
defaultLng: process.env.SHARELATEX_SITE_LANGUAGE || "en"
|
||||
},
|
||||
defaultLng: process.env.SHARELATEX_SITE_LANGUAGE || 'en',
|
||||
},
|
||||
|
||||
currentImageName: process.env.TEX_LIVE_DOCKER_IMAGE,
|
||||
|
||||
apis: {
|
||||
web: {
|
||||
url: "http://localhost:3000",
|
||||
url: 'http://localhost:3000',
|
||||
user: httpAuthUser,
|
||||
pass: httpAuthPass
|
||||
pass: httpAuthPass,
|
||||
},
|
||||
project_history: {
|
||||
enabled: false
|
||||
}
|
||||
enabled: false,
|
||||
},
|
||||
},
|
||||
references: {},
|
||||
notifications: undefined,
|
||||
|
@ -259,36 +301,36 @@ const settings = {
|
|||
dropbox: true,
|
||||
versioning: true,
|
||||
compileTimeout: parseIntOrFail(process.env.COMPILE_TIMEOUT || 180),
|
||||
compileGroup: "standard",
|
||||
compileGroup: 'standard',
|
||||
trackChanges: true,
|
||||
templates: true,
|
||||
references: true
|
||||
references: true,
|
||||
},
|
||||
}
|
||||
};
|
||||
|
||||
// # OPTIONAL CONFIGURABLE SETTINGS
|
||||
|
||||
if (process.env.SHARELATEX_LEFT_FOOTER != null) {
|
||||
try {
|
||||
settings.nav.left_footer = JSON.parse(process.env.SHARELATEX_LEFT_FOOTER);
|
||||
settings.nav.left_footer = JSON.parse(process.env.SHARELATEX_LEFT_FOOTER)
|
||||
} catch (error) {
|
||||
e = error;
|
||||
console.error("could not parse SHARELATEX_LEFT_FOOTER, not valid JSON");
|
||||
e = error
|
||||
console.error('could not parse SHARELATEX_LEFT_FOOTER, not valid JSON')
|
||||
}
|
||||
}
|
||||
|
||||
if (process.env.SHARELATEX_RIGHT_FOOTER != null) {
|
||||
settings.nav.right_footer = process.env.SHARELATEX_RIGHT_FOOTER;
|
||||
settings.nav.right_footer = process.env.SHARELATEX_RIGHT_FOOTER
|
||||
try {
|
||||
settings.nav.right_footer = JSON.parse(process.env.SHARELATEX_RIGHT_FOOTER);
|
||||
settings.nav.right_footer = JSON.parse(process.env.SHARELATEX_RIGHT_FOOTER)
|
||||
} catch (error1) {
|
||||
e = error1;
|
||||
console.error("could not parse SHARELATEX_RIGHT_FOOTER, not valid JSON");
|
||||
e = error1
|
||||
console.error('could not parse SHARELATEX_RIGHT_FOOTER, not valid JSON')
|
||||
}
|
||||
}
|
||||
|
||||
if (process.env.SHARELATEX_HEADER_IMAGE_URL != null) {
|
||||
settings.nav.custom_logo = process.env.SHARELATEX_HEADER_IMAGE_URL;
|
||||
settings.nav.custom_logo = process.env.SHARELATEX_HEADER_IMAGE_URL
|
||||
}
|
||||
|
||||
if (process.env.SHARELATEX_HEADER_NAV_LINKS != null) {
|
||||
|
@ -299,21 +341,20 @@ if (process.env.SHARELATEX_HEADER_NAV_LINKS != null) {
|
|||
# See https://github.com/sharelatex/sharelatex/wiki/Configuring-Headers,-Footers-&-Logo
|
||||
#
|
||||
# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # #\
|
||||
`
|
||||
);
|
||||
`)
|
||||
}
|
||||
|
||||
if (process.env.SHARELATEX_HEADER_EXTRAS != null) {
|
||||
try {
|
||||
settings.nav.header_extras = JSON.parse(process.env.SHARELATEX_HEADER_EXTRAS);
|
||||
settings.nav.header_extras = JSON.parse(
|
||||
process.env.SHARELATEX_HEADER_EXTRAS
|
||||
)
|
||||
} catch (error2) {
|
||||
e = error2;
|
||||
console.error("could not parse SHARELATEX_HEADER_EXTRAS, not valid JSON");
|
||||
e = error2
|
||||
console.error('could not parse SHARELATEX_HEADER_EXTRAS, not valid JSON')
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
// Sending Email
|
||||
// -------------
|
||||
//
|
||||
|
@ -323,12 +364,10 @@ if (process.env.SHARELATEX_HEADER_EXTRAS != null) {
|
|||
//
|
||||
// http://www.nodemailer.com/docs/transports
|
||||
|
||||
|
||||
if (process.env.SHARELATEX_EMAIL_FROM_ADDRESS != null) {
|
||||
|
||||
settings.email = {
|
||||
fromAddress: process.env.SHARELATEX_EMAIL_FROM_ADDRESS,
|
||||
replyTo: process.env.SHARELATEX_EMAIL_REPLY_TO || "",
|
||||
replyTo: process.env.SHARELATEX_EMAIL_REPLY_TO || '',
|
||||
driver: process.env.SHARELATEX_EMAIL_DRIVER,
|
||||
parameters: {
|
||||
// AWS Creds
|
||||
|
@ -341,65 +380,73 @@ if (process.env.SHARELATEX_EMAIL_FROM_ADDRESS != null) {
|
|||
secure: parse(process.env.SHARELATEX_EMAIL_SMTP_SECURE),
|
||||
ignoreTLS: parse(process.env.SHARELATEX_EMAIL_SMTP_IGNORE_TLS),
|
||||
name: process.env.SHARELATEX_EMAIL_SMTP_NAME,
|
||||
logger: process.env.SHARELATEX_EMAIL_SMTP_LOGGER === 'true'
|
||||
logger: process.env.SHARELATEX_EMAIL_SMTP_LOGGER === 'true',
|
||||
},
|
||||
|
||||
textEncoding: process.env.SHARELATEX_EMAIL_TEXT_ENCODING,
|
||||
template: {
|
||||
customFooter: process.env.SHARELATEX_CUSTOM_EMAIL_FOOTER
|
||||
customFooter: process.env.SHARELATEX_CUSTOM_EMAIL_FOOTER,
|
||||
},
|
||||
}
|
||||
};
|
||||
|
||||
if (process.env.SHARELATEX_EMAIL_AWS_SES_REGION != null) {
|
||||
settings.email.parameters.region = process.env.SHARELATEX_EMAIL_AWS_SES_REGION;
|
||||
settings.email.parameters.region =
|
||||
process.env.SHARELATEX_EMAIL_AWS_SES_REGION
|
||||
}
|
||||
|
||||
if ((process.env.SHARELATEX_EMAIL_SMTP_USER != null) || (process.env.SHARELATEX_EMAIL_SMTP_PASS != null)) {
|
||||
if (
|
||||
process.env.SHARELATEX_EMAIL_SMTP_USER != null ||
|
||||
process.env.SHARELATEX_EMAIL_SMTP_PASS != null
|
||||
) {
|
||||
settings.email.parameters.auth = {
|
||||
user: process.env.SHARELATEX_EMAIL_SMTP_USER,
|
||||
pass: process.env.SHARELATEX_EMAIL_SMTP_PASS
|
||||
};
|
||||
pass: process.env.SHARELATEX_EMAIL_SMTP_PASS,
|
||||
}
|
||||
}
|
||||
|
||||
if (process.env.SHARELATEX_EMAIL_SMTP_TLS_REJECT_UNAUTH != null) {
|
||||
settings.email.parameters.tls =
|
||||
{rejectUnauthorized: parse(process.env.SHARELATEX_EMAIL_SMTP_TLS_REJECT_UNAUTH)};
|
||||
settings.email.parameters.tls = {
|
||||
rejectUnauthorized: parse(
|
||||
process.env.SHARELATEX_EMAIL_SMTP_TLS_REJECT_UNAUTH
|
||||
),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// i18n
|
||||
if (process.env.SHARELATEX_LANG_DOMAIN_MAPPING != null) {
|
||||
|
||||
settings.i18n.subdomainLang = parse(process.env.SHARELATEX_LANG_DOMAIN_MAPPING);
|
||||
settings.i18n.subdomainLang = parse(
|
||||
process.env.SHARELATEX_LANG_DOMAIN_MAPPING
|
||||
)
|
||||
}
|
||||
|
||||
// Password Settings
|
||||
// -----------
|
||||
// These restrict the passwords users can use when registering
|
||||
// opts are from http://antelle.github.io/passfield
|
||||
if (process.env.SHARELATEX_PASSWORD_VALIDATION_PATTERN || process.env.SHARELATEX_PASSWORD_VALIDATION_MIN_LENGTH || process.env.SHARELATEX_PASSWORD_VALIDATION_MAX_LENGTH) {
|
||||
|
||||
if (
|
||||
process.env.SHARELATEX_PASSWORD_VALIDATION_PATTERN ||
|
||||
process.env.SHARELATEX_PASSWORD_VALIDATION_MIN_LENGTH ||
|
||||
process.env.SHARELATEX_PASSWORD_VALIDATION_MAX_LENGTH
|
||||
) {
|
||||
settings.passwordStrengthOptions = {
|
||||
pattern: process.env.SHARELATEX_PASSWORD_VALIDATION_PATTERN || "aA$3",
|
||||
length: {min:process.env.SHARELATEX_PASSWORD_VALIDATION_MIN_LENGTH || 8, max: process.env.SHARELATEX_PASSWORD_VALIDATION_MAX_LENGTH || 150}
|
||||
};
|
||||
pattern: process.env.SHARELATEX_PASSWORD_VALIDATION_PATTERN || 'aA$3',
|
||||
length: {
|
||||
min: process.env.SHARELATEX_PASSWORD_VALIDATION_MIN_LENGTH || 8,
|
||||
max: process.env.SHARELATEX_PASSWORD_VALIDATION_MAX_LENGTH || 150,
|
||||
},
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
||||
// ######################
|
||||
// ShareLaTeX Server Pro
|
||||
// ######################
|
||||
|
||||
if (parse(process.env.SHARELATEX_IS_SERVER_PRO) === true) {
|
||||
settings.bypassPercentageRollouts = true;
|
||||
settings.apis.references =
|
||||
{url: "http://localhost:3040"};
|
||||
settings.bypassPercentageRollouts = true
|
||||
settings.apis.references = { url: 'http://localhost:3040' }
|
||||
}
|
||||
|
||||
|
||||
// LDAP - SERVER PRO ONLY
|
||||
// ----------
|
||||
|
||||
|
@ -411,18 +458,21 @@ if (process.env.SHARELATEX_LDAP_HOST) {
|
|||
# See https://github.com/sharelatex/sharelatex/wiki/Server-Pro:-LDAP-Config
|
||||
#
|
||||
# # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # # #\
|
||||
`
|
||||
);
|
||||
`)
|
||||
}
|
||||
|
||||
if (process.env.SHARELATEX_LDAP_URL) {
|
||||
let _ldap_connect_timeout, _ldap_group_search_attribs, _ldap_search_attribs, _ldap_timeout;
|
||||
settings.externalAuth = true;
|
||||
let _ldap_connect_timeout,
|
||||
_ldap_group_search_attribs,
|
||||
_ldap_search_attribs,
|
||||
_ldap_timeout
|
||||
settings.externalAuth = true
|
||||
settings.ldap = {
|
||||
emailAtt: process.env.SHARELATEX_LDAP_EMAIL_ATT,
|
||||
nameAtt: process.env.SHARELATEX_LDAP_NAME_ATT,
|
||||
lastNameAtt: process.env.SHARELATEX_LDAP_LAST_NAME_ATT,
|
||||
updateUserDetailsOnLogin: process.env.SHARELATEX_LDAP_UPDATE_USER_DETAILS_ON_LOGIN === 'true',
|
||||
updateUserDetailsOnLogin:
|
||||
process.env.SHARELATEX_LDAP_UPDATE_USER_DETAILS_ON_LOGIN === 'true',
|
||||
placeholder: process.env.SHARELATEX_LDAP_PLACEHOLDER,
|
||||
server: {
|
||||
url: process.env.SHARELATEX_LDAP_URL,
|
||||
|
@ -432,94 +482,108 @@ if (process.env.SHARELATEX_LDAP_URL) {
|
|||
searchBase: process.env.SHARELATEX_LDAP_SEARCH_BASE,
|
||||
searchScope: process.env.SHARELATEX_LDAP_SEARCH_SCOPE,
|
||||
searchFilter: process.env.SHARELATEX_LDAP_SEARCH_FILTER,
|
||||
searchAttributes: (
|
||||
(_ldap_search_attribs = process.env.SHARELATEX_LDAP_SEARCH_ATTRIBUTES) ?
|
||||
(() => { try {
|
||||
return JSON.parse(_ldap_search_attribs);
|
||||
searchAttributes: (_ldap_search_attribs =
|
||||
process.env.SHARELATEX_LDAP_SEARCH_ATTRIBUTES)
|
||||
? (() => {
|
||||
try {
|
||||
return JSON.parse(_ldap_search_attribs)
|
||||
} catch (error3) {
|
||||
e = error3;
|
||||
return console.error("could not parse SHARELATEX_LDAP_SEARCH_ATTRIBUTES");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
),
|
||||
e = error3
|
||||
return console.error(
|
||||
'could not parse SHARELATEX_LDAP_SEARCH_ATTRIBUTES'
|
||||
)
|
||||
}
|
||||
})()
|
||||
: undefined,
|
||||
groupDnProperty: process.env.SHARELATEX_LDAP_GROUP_DN_PROPERTY,
|
||||
groupSearchBase: process.env.SHARELATEX_LDAP_GROUP_SEARCH_BASE,
|
||||
groupSearchScope: process.env.SHARELATEX_LDAP_GROUP_SEARCH_SCOPE,
|
||||
groupSearchFilter: process.env.SHARELATEX_LDAP_GROUP_SEARCH_FILTER,
|
||||
groupSearchAttributes: (
|
||||
(_ldap_group_search_attribs = process.env.SHARELATEX_LDAP_GROUP_SEARCH_ATTRIBUTES) ?
|
||||
(() => { try {
|
||||
return JSON.parse(_ldap_group_search_attribs);
|
||||
groupSearchAttributes: (_ldap_group_search_attribs =
|
||||
process.env.SHARELATEX_LDAP_GROUP_SEARCH_ATTRIBUTES)
|
||||
? (() => {
|
||||
try {
|
||||
return JSON.parse(_ldap_group_search_attribs)
|
||||
} catch (error4) {
|
||||
e = error4;
|
||||
return console.error("could not parse SHARELATEX_LDAP_GROUP_SEARCH_ATTRIBUTES");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
),
|
||||
cache: process.env.SHARELATEX_LDAP_CACHE === 'true',
|
||||
timeout: (
|
||||
(_ldap_timeout = process.env.SHARELATEX_LDAP_TIMEOUT) ?
|
||||
(() => { try {
|
||||
return parseIntOrFail(_ldap_timeout);
|
||||
} catch (error5) {
|
||||
e = error5;
|
||||
return console.error("Cannot parse SHARELATEX_LDAP_TIMEOUT");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
),
|
||||
connectTimeout: (
|
||||
(_ldap_connect_timeout = process.env.SHARELATEX_LDAP_CONNECT_TIMEOUT) ?
|
||||
(() => { try {
|
||||
return parseIntOrFail(_ldap_connect_timeout);
|
||||
} catch (error6) {
|
||||
e = error6;
|
||||
return console.error("Cannot parse SHARELATEX_LDAP_CONNECT_TIMEOUT");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
e = error4
|
||||
return console.error(
|
||||
'could not parse SHARELATEX_LDAP_GROUP_SEARCH_ATTRIBUTES'
|
||||
)
|
||||
}
|
||||
};
|
||||
|
||||
if (process.env.SHARELATEX_LDAP_TLS_OPTS_CA_PATH) {
|
||||
let ca, ca_paths;
|
||||
})()
|
||||
: undefined,
|
||||
cache: process.env.SHARELATEX_LDAP_CACHE === 'true',
|
||||
timeout: (_ldap_timeout = process.env.SHARELATEX_LDAP_TIMEOUT)
|
||||
? (() => {
|
||||
try {
|
||||
ca = JSON.parse(process.env.SHARELATEX_LDAP_TLS_OPTS_CA_PATH);
|
||||
} catch (error7) {
|
||||
e = error7;
|
||||
console.error("could not parse SHARELATEX_LDAP_TLS_OPTS_CA_PATH, invalid JSON");
|
||||
return parseIntOrFail(_ldap_timeout)
|
||||
} catch (error5) {
|
||||
e = error5
|
||||
return console.error('Cannot parse SHARELATEX_LDAP_TIMEOUT')
|
||||
}
|
||||
})()
|
||||
: undefined,
|
||||
connectTimeout: (_ldap_connect_timeout =
|
||||
process.env.SHARELATEX_LDAP_CONNECT_TIMEOUT)
|
||||
? (() => {
|
||||
try {
|
||||
return parseIntOrFail(_ldap_connect_timeout)
|
||||
} catch (error6) {
|
||||
e = error6
|
||||
return console.error(
|
||||
'Cannot parse SHARELATEX_LDAP_CONNECT_TIMEOUT'
|
||||
)
|
||||
}
|
||||
})()
|
||||
: undefined,
|
||||
},
|
||||
}
|
||||
|
||||
if (typeof(ca) === 'string') {
|
||||
ca_paths = [ca];
|
||||
} else if ((typeof(ca) === 'object') && ((ca != null ? ca.length : undefined) != null)) {
|
||||
ca_paths = ca;
|
||||
if (process.env.SHARELATEX_LDAP_TLS_OPTS_CA_PATH) {
|
||||
let ca, ca_paths
|
||||
try {
|
||||
ca = JSON.parse(process.env.SHARELATEX_LDAP_TLS_OPTS_CA_PATH)
|
||||
} catch (error7) {
|
||||
e = error7
|
||||
console.error(
|
||||
'could not parse SHARELATEX_LDAP_TLS_OPTS_CA_PATH, invalid JSON'
|
||||
)
|
||||
}
|
||||
|
||||
if (typeof ca === 'string') {
|
||||
ca_paths = [ca]
|
||||
} else if (
|
||||
typeof ca === 'object' &&
|
||||
(ca != null ? ca.length : undefined) != null
|
||||
) {
|
||||
ca_paths = ca
|
||||
} else {
|
||||
console.error("problem parsing SHARELATEX_LDAP_TLS_OPTS_CA_PATH");
|
||||
console.error('problem parsing SHARELATEX_LDAP_TLS_OPTS_CA_PATH')
|
||||
}
|
||||
|
||||
settings.ldap.server.tlsOptions = {
|
||||
rejectUnauthorized: process.env.SHARELATEX_LDAP_TLS_OPTS_REJECT_UNAUTH === "true",
|
||||
ca:ca_paths // e.g.'/etc/ldap/ca_certs.pem'
|
||||
};
|
||||
rejectUnauthorized:
|
||||
process.env.SHARELATEX_LDAP_TLS_OPTS_REJECT_UNAUTH === 'true',
|
||||
ca: ca_paths, // e.g.'/etc/ldap/ca_certs.pem'
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
if (process.env.SHARELATEX_SAML_ENTRYPOINT) {
|
||||
// NOTE: see https://github.com/node-saml/passport-saml/blob/master/README.md for docs of `server` options
|
||||
let _saml_additionalAuthorizeParams, _saml_additionalLogoutParams, _saml_additionalParams, _saml_expiration, _saml_skew;
|
||||
settings.externalAuth = true;
|
||||
let _saml_additionalAuthorizeParams,
|
||||
_saml_additionalLogoutParams,
|
||||
_saml_additionalParams,
|
||||
_saml_expiration,
|
||||
_saml_skew
|
||||
settings.externalAuth = true
|
||||
settings.saml = {
|
||||
updateUserDetailsOnLogin: process.env.SHARELATEX_SAML_UPDATE_USER_DETAILS_ON_LOGIN === 'true',
|
||||
updateUserDetailsOnLogin:
|
||||
process.env.SHARELATEX_SAML_UPDATE_USER_DETAILS_ON_LOGIN === 'true',
|
||||
identityServiceName: process.env.SHARELATEX_SAML_IDENTITY_SERVICE_NAME,
|
||||
emailField: process.env.SHARELATEX_SAML_EMAIL_FIELD || process.env.SHARELATEX_SAML_EMAIL_FIELD_NAME,
|
||||
emailField:
|
||||
process.env.SHARELATEX_SAML_EMAIL_FIELD ||
|
||||
process.env.SHARELATEX_SAML_EMAIL_FIELD_NAME,
|
||||
firstNameField: process.env.SHARELATEX_SAML_FIRST_NAME_FIELD,
|
||||
lastNameField: process.env.SHARELATEX_SAML_LAST_NAME_FIELD,
|
||||
server: {
|
||||
|
@ -531,162 +595,184 @@ if (process.env.SHARELATEX_SAML_ENTRYPOINT) {
|
|||
decryptionCert: process.env.SHARELATEX_SAML_DECRYPTION_CERT,
|
||||
signatureAlgorithm: process.env.SHARELATEX_SAML_SIGNATURE_ALGORITHM,
|
||||
identifierFormat: process.env.SHARELATEX_SAML_IDENTIFIER_FORMAT,
|
||||
attributeConsumingServiceIndex: process.env.SHARELATEX_SAML_ATTRIBUTE_CONSUMING_SERVICE_INDEX,
|
||||
attributeConsumingServiceIndex:
|
||||
process.env.SHARELATEX_SAML_ATTRIBUTE_CONSUMING_SERVICE_INDEX,
|
||||
authnContext: process.env.SHARELATEX_SAML_AUTHN_CONTEXT,
|
||||
authnRequestBinding: process.env.SHARELATEX_SAML_AUTHN_REQUEST_BINDING,
|
||||
validateInResponseTo: process.env.SHARELATEX_SAML_VALIDATE_IN_RESPONSE_TO,
|
||||
cacheProvider: process.env.SHARELATEX_SAML_CACHE_PROVIDER,
|
||||
logoutUrl: process.env.SHARELATEX_SAML_LOGOUT_URL,
|
||||
logoutCallbackUrl: process.env.SHARELATEX_SAML_LOGOUT_CALLBACK_URL,
|
||||
disableRequestedAuthnContext: process.env.SHARELATEX_SAML_DISABLE_REQUESTED_AUTHN_CONTEXT === 'true',
|
||||
disableRequestedAuthnContext:
|
||||
process.env.SHARELATEX_SAML_DISABLE_REQUESTED_AUTHN_CONTEXT === 'true',
|
||||
forceAuthn: process.env.SHARELATEX_SAML_FORCE_AUTHN === 'true',
|
||||
skipRequestCompression: process.env.SHARELATEX_SAML_SKIP_REQUEST_COMPRESSION === 'true',
|
||||
acceptedClockSkewMs: (
|
||||
(_saml_skew = process.env.SHARELATEX_SAML_ACCEPTED_CLOCK_SKEW_MS) ?
|
||||
(() => { try {
|
||||
return parseIntOrFail(_saml_skew);
|
||||
skipRequestCompression:
|
||||
process.env.SHARELATEX_SAML_SKIP_REQUEST_COMPRESSION === 'true',
|
||||
acceptedClockSkewMs: (_saml_skew =
|
||||
process.env.SHARELATEX_SAML_ACCEPTED_CLOCK_SKEW_MS)
|
||||
? (() => {
|
||||
try {
|
||||
return parseIntOrFail(_saml_skew)
|
||||
} catch (error8) {
|
||||
e = error8;
|
||||
return console.error("Cannot parse SHARELATEX_SAML_ACCEPTED_CLOCK_SKEW_MS");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
),
|
||||
requestIdExpirationPeriodMs: (
|
||||
(_saml_expiration = process.env.SHARELATEX_SAML_REQUEST_ID_EXPIRATION_PERIOD_MS) ?
|
||||
(() => { try {
|
||||
return parseIntOrFail(_saml_expiration);
|
||||
} catch (error9) {
|
||||
e = error9;
|
||||
return console.error("Cannot parse SHARELATEX_SAML_REQUEST_ID_EXPIRATION_PERIOD_MS");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
),
|
||||
additionalParams: (
|
||||
(_saml_additionalParams = process.env.SHARELATEX_SAML_ADDITIONAL_PARAMS) ?
|
||||
(() => { try {
|
||||
return JSON.parse(_saml_additionalParams);
|
||||
} catch (error10) {
|
||||
e = error10;
|
||||
return console.error("Cannot parse SHARELATEX_SAML_ADDITIONAL_PARAMS");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
),
|
||||
additionalAuthorizeParams: (
|
||||
(_saml_additionalAuthorizeParams = process.env.SHARELATEX_SAML_ADDITIONAL_AUTHORIZE_PARAMS) ?
|
||||
(() => { try {
|
||||
return JSON.parse(_saml_additionalAuthorizeParams );
|
||||
} catch (error11) {
|
||||
e = error11;
|
||||
return console.error("Cannot parse SHARELATEX_SAML_ADDITIONAL_AUTHORIZE_PARAMS");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
),
|
||||
additionalLogoutParams: (
|
||||
(_saml_additionalLogoutParams = process.env.SHARELATEX_SAML_ADDITIONAL_LOGOUT_PARAMS) ?
|
||||
(() => { try {
|
||||
return JSON.parse(_saml_additionalLogoutParams );
|
||||
} catch (error12) {
|
||||
e = error12;
|
||||
return console.error("Cannot parse SHARELATEX_SAML_ADDITIONAL_LOGOUT_PARAMS");
|
||||
} })()
|
||||
:
|
||||
undefined
|
||||
e = error8
|
||||
return console.error(
|
||||
'Cannot parse SHARELATEX_SAML_ACCEPTED_CLOCK_SKEW_MS'
|
||||
)
|
||||
}
|
||||
};
|
||||
})()
|
||||
: undefined,
|
||||
requestIdExpirationPeriodMs: (_saml_expiration =
|
||||
process.env.SHARELATEX_SAML_REQUEST_ID_EXPIRATION_PERIOD_MS)
|
||||
? (() => {
|
||||
try {
|
||||
return parseIntOrFail(_saml_expiration)
|
||||
} catch (error9) {
|
||||
e = error9
|
||||
return console.error(
|
||||
'Cannot parse SHARELATEX_SAML_REQUEST_ID_EXPIRATION_PERIOD_MS'
|
||||
)
|
||||
}
|
||||
})()
|
||||
: undefined,
|
||||
additionalParams: (_saml_additionalParams =
|
||||
process.env.SHARELATEX_SAML_ADDITIONAL_PARAMS)
|
||||
? (() => {
|
||||
try {
|
||||
return JSON.parse(_saml_additionalParams)
|
||||
} catch (error10) {
|
||||
e = error10
|
||||
return console.error(
|
||||
'Cannot parse SHARELATEX_SAML_ADDITIONAL_PARAMS'
|
||||
)
|
||||
}
|
||||
})()
|
||||
: undefined,
|
||||
additionalAuthorizeParams: (_saml_additionalAuthorizeParams =
|
||||
process.env.SHARELATEX_SAML_ADDITIONAL_AUTHORIZE_PARAMS)
|
||||
? (() => {
|
||||
try {
|
||||
return JSON.parse(_saml_additionalAuthorizeParams)
|
||||
} catch (error11) {
|
||||
e = error11
|
||||
return console.error(
|
||||
'Cannot parse SHARELATEX_SAML_ADDITIONAL_AUTHORIZE_PARAMS'
|
||||
)
|
||||
}
|
||||
})()
|
||||
: undefined,
|
||||
additionalLogoutParams: (_saml_additionalLogoutParams =
|
||||
process.env.SHARELATEX_SAML_ADDITIONAL_LOGOUT_PARAMS)
|
||||
? (() => {
|
||||
try {
|
||||
return JSON.parse(_saml_additionalLogoutParams)
|
||||
} catch (error12) {
|
||||
e = error12
|
||||
return console.error(
|
||||
'Cannot parse SHARELATEX_SAML_ADDITIONAL_LOGOUT_PARAMS'
|
||||
)
|
||||
}
|
||||
})()
|
||||
: undefined,
|
||||
},
|
||||
}
|
||||
|
||||
// SHARELATEX_SAML_CERT cannot be empty
|
||||
// https://github.com/node-saml/passport-saml/commit/f6b1c885c0717f1083c664345556b535f217c102
|
||||
if (process.env.SHARELATEX_SAML_CERT) {
|
||||
settings.saml.server.cert = process.env.SHARELATEX_SAML_CERT;
|
||||
settings.saml.server.privateCert = process.env.SHARELATEX_SAML_PRIVATE_CERT;
|
||||
settings.saml.server.cert = process.env.SHARELATEX_SAML_CERT
|
||||
settings.saml.server.privateCert = process.env.SHARELATEX_SAML_PRIVATE_CERT
|
||||
}
|
||||
}
|
||||
|
||||
// Compiler
|
||||
// --------
|
||||
if (process.env.SANDBOXED_COMPILES === "true") {
|
||||
if (process.env.SANDBOXED_COMPILES === 'true') {
|
||||
settings.clsi = {
|
||||
dockerRunner: true,
|
||||
docker: {
|
||||
image: process.env.TEX_LIVE_DOCKER_IMAGE,
|
||||
env: {
|
||||
HOME: "/tmp",
|
||||
PATH: process.env.COMPILER_PATH || "/usr/local/texlive/2015/bin/x86_64-linux:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"
|
||||
HOME: '/tmp',
|
||||
PATH:
|
||||
process.env.COMPILER_PATH ||
|
||||
'/usr/local/texlive/2015/bin/x86_64-linux:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin',
|
||||
},
|
||||
user: 'www-data',
|
||||
},
|
||||
user: "www-data"
|
||||
}
|
||||
};
|
||||
|
||||
if ((settings.path == null)) {
|
||||
settings.path = {};
|
||||
if (settings.path == null) {
|
||||
settings.path = {}
|
||||
}
|
||||
settings.path.synctexBaseDir = () => "/compile";
|
||||
settings.path.synctexBaseDir = () => '/compile'
|
||||
if (process.env.SANDBOXED_COMPILES_SIBLING_CONTAINERS === 'true') {
|
||||
console.log("Using sibling containers for sandboxed compiles");
|
||||
console.log('Using sibling containers for sandboxed compiles')
|
||||
if (process.env.SANDBOXED_COMPILES_HOST_DIR) {
|
||||
settings.path.sandboxedCompilesHostDir = process.env.SANDBOXED_COMPILES_HOST_DIR;
|
||||
settings.path.sandboxedCompilesHostDir =
|
||||
process.env.SANDBOXED_COMPILES_HOST_DIR
|
||||
} else {
|
||||
console.error('Sibling containers, but SANDBOXED_COMPILES_HOST_DIR not set');
|
||||
console.error(
|
||||
'Sibling containers, but SANDBOXED_COMPILES_HOST_DIR not set'
|
||||
)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
// Templates
|
||||
// ---------
|
||||
if (process.env.SHARELATEX_TEMPLATES_USER_ID) {
|
||||
settings.templates = {
|
||||
mountPointUrl: "/templates",
|
||||
user_id: process.env.SHARELATEX_TEMPLATES_USER_ID
|
||||
};
|
||||
|
||||
settings.templateLinks = parse(process.env.SHARELATEX_NEW_PROJECT_TEMPLATE_LINKS);
|
||||
mountPointUrl: '/templates',
|
||||
user_id: process.env.SHARELATEX_TEMPLATES_USER_ID,
|
||||
}
|
||||
|
||||
settings.templateLinks = parse(
|
||||
process.env.SHARELATEX_NEW_PROJECT_TEMPLATE_LINKS
|
||||
)
|
||||
}
|
||||
|
||||
// /Learn
|
||||
// -------
|
||||
if (process.env.SHARELATEX_PROXY_LEARN != null) {
|
||||
settings.proxyLearn = parse(process.env.SHARELATEX_PROXY_LEARN);
|
||||
settings.proxyLearn = parse(process.env.SHARELATEX_PROXY_LEARN)
|
||||
}
|
||||
|
||||
|
||||
// /References
|
||||
// -----------
|
||||
if (process.env.SHARELATEX_ELASTICSEARCH_URL != null) {
|
||||
settings.references.elasticsearch =
|
||||
{host: process.env.SHARELATEX_ELASTICSEARCH_URL};
|
||||
settings.references.elasticsearch = {
|
||||
host: process.env.SHARELATEX_ELASTICSEARCH_URL,
|
||||
}
|
||||
}
|
||||
|
||||
// TeX Live Images
|
||||
// -----------
|
||||
if (process.env.ALL_TEX_LIVE_DOCKER_IMAGES != null) {
|
||||
allTexLiveDockerImages = process.env.ALL_TEX_LIVE_DOCKER_IMAGES.split(',');
|
||||
allTexLiveDockerImages = process.env.ALL_TEX_LIVE_DOCKER_IMAGES.split(',')
|
||||
}
|
||||
if (process.env.ALL_TEX_LIVE_DOCKER_IMAGE_NAMES != null) {
|
||||
allTexLiveDockerImageNames = process.env.ALL_TEX_LIVE_DOCKER_IMAGE_NAMES.split(',');
|
||||
allTexLiveDockerImageNames =
|
||||
process.env.ALL_TEX_LIVE_DOCKER_IMAGE_NAMES.split(',')
|
||||
}
|
||||
if (allTexLiveDockerImages != null) {
|
||||
settings.allowedImageNames = [];
|
||||
settings.allowedImageNames = []
|
||||
for (let index = 0; index < allTexLiveDockerImages.length; index++) {
|
||||
const fullImageName = allTexLiveDockerImages[index];
|
||||
const imageName = Path.basename(fullImageName);
|
||||
const imageDesc = (allTexLiveDockerImageNames != null) ? allTexLiveDockerImageNames[index] : imageName;
|
||||
settings.allowedImageNames.push({ imageName, imageDesc });
|
||||
const fullImageName = allTexLiveDockerImages[index]
|
||||
const imageName = Path.basename(fullImageName)
|
||||
const imageDesc =
|
||||
allTexLiveDockerImageNames != null
|
||||
? allTexLiveDockerImageNames[index]
|
||||
: imageName
|
||||
settings.allowedImageNames.push({ imageName, imageDesc })
|
||||
}
|
||||
}
|
||||
|
||||
// With lots of incoming and outgoing HTTP connections to different services,
|
||||
// sometimes long running, it is a good idea to increase the default number
|
||||
// of sockets that Node will hold open.
|
||||
const http = require('http');
|
||||
http.globalAgent.maxSockets = 300;
|
||||
const https = require('https');
|
||||
https.globalAgent.maxSockets = 300;
|
||||
const http = require('http')
|
||||
http.globalAgent.maxSockets = 300
|
||||
const https = require('https')
|
||||
https.globalAgent.maxSockets = 300
|
||||
|
||||
module.exports = settings;
|
||||
module.exports = settings
|
||||
|
|
|
@ -12,34 +12,54 @@
|
|||
*/
|
||||
|
||||
module.exports = function (grunt) {
|
||||
|
||||
grunt.registerTask('user:create-admin', "Create a user with the given email address and make them an admin. Update in place if the user already exists. Usage: grunt user:create-admin --email joe@example.com", function() {
|
||||
const done = this.async();
|
||||
const email = grunt.option("email");
|
||||
if ((email == null)) {
|
||||
console.error("Usage: grunt user:create-admin --email=joe@example.com");
|
||||
process.exit(1);
|
||||
grunt.registerTask(
|
||||
'user:create-admin',
|
||||
'Create a user with the given email address and make them an admin. Update in place if the user already exists. Usage: grunt user:create-admin --email joe@example.com',
|
||||
function () {
|
||||
const done = this.async()
|
||||
const email = grunt.option('email')
|
||||
if (email == null) {
|
||||
console.error('Usage: grunt user:create-admin --email=joe@example.com')
|
||||
process.exit(1)
|
||||
}
|
||||
|
||||
const settings = require("settings-sharelatex");
|
||||
const mongodb = require("../web/app/src/infrastructure/mongodb");
|
||||
const UserRegistrationHandler = require("../web/app/src/Features/User/UserRegistrationHandler");
|
||||
const OneTimeTokenHandler = require("../web/app/src/Features/Security/OneTimeTokenHandler");
|
||||
return mongodb.waitForDb().then(() => UserRegistrationHandler.registerNewUser({
|
||||
const settings = require('settings-sharelatex')
|
||||
const mongodb = require('../web/app/src/infrastructure/mongodb')
|
||||
const UserRegistrationHandler = require('../web/app/src/Features/User/UserRegistrationHandler')
|
||||
const OneTimeTokenHandler = require('../web/app/src/Features/Security/OneTimeTokenHandler')
|
||||
return mongodb.waitForDb().then(() =>
|
||||
UserRegistrationHandler.registerNewUser(
|
||||
{
|
||||
email,
|
||||
password: require("crypto").randomBytes(32).toString("hex")
|
||||
}, function(error, user) {
|
||||
if ((error != null) && ((error != null ? error.message : undefined) !== "EmailAlreadyRegistered")) {
|
||||
throw error;
|
||||
password: require('crypto').randomBytes(32).toString('hex'),
|
||||
},
|
||||
function (error, user) {
|
||||
if (
|
||||
error != null &&
|
||||
(error != null ? error.message : undefined) !==
|
||||
'EmailAlreadyRegistered'
|
||||
) {
|
||||
throw error
|
||||
}
|
||||
user.isAdmin = true;
|
||||
user.isAdmin = true
|
||||
return user.save(function (error) {
|
||||
if (error != null) { throw error; }
|
||||
const ONE_WEEK = 7 * 24 * 60 * 60; // seconds
|
||||
return OneTimeTokenHandler.getNewToken("password", { expiresIn: ONE_WEEK, email:user.email, user_id: user._id.toString() }, function(err, token){
|
||||
if (err != null) { return next(err); }
|
||||
if (error != null) {
|
||||
throw error
|
||||
}
|
||||
const ONE_WEEK = 7 * 24 * 60 * 60 // seconds
|
||||
return OneTimeTokenHandler.getNewToken(
|
||||
'password',
|
||||
{
|
||||
expiresIn: ONE_WEEK,
|
||||
email: user.email,
|
||||
user_id: user._id.toString(),
|
||||
},
|
||||
function (err, token) {
|
||||
if (err != null) {
|
||||
return next(err)
|
||||
}
|
||||
|
||||
console.log("");
|
||||
console.log('')
|
||||
console.log(`\
|
||||
Successfully created ${email} as an admin user.
|
||||
|
||||
|
@ -47,39 +67,50 @@ Please visit the following URL to set a password for ${email} and log in:
|
|||
|
||||
${settings.siteUrl}/user/password/set?passwordResetToken=${token}
|
||||
\
|
||||
`
|
||||
);
|
||||
return done();
|
||||
});
|
||||
});
|
||||
}));
|
||||
});
|
||||
`)
|
||||
return done()
|
||||
}
|
||||
)
|
||||
})
|
||||
}
|
||||
)
|
||||
)
|
||||
}
|
||||
)
|
||||
|
||||
return grunt.registerTask('user:delete', "deletes a user and all their data, Usage: grunt user:delete --email joe@example.com", function() {
|
||||
const done = this.async();
|
||||
const email = grunt.option("email");
|
||||
if ((email == null)) {
|
||||
console.error("Usage: grunt user:delete --email=joe@example.com");
|
||||
process.exit(1);
|
||||
return grunt.registerTask(
|
||||
'user:delete',
|
||||
'deletes a user and all their data, Usage: grunt user:delete --email joe@example.com',
|
||||
function () {
|
||||
const done = this.async()
|
||||
const email = grunt.option('email')
|
||||
if (email == null) {
|
||||
console.error('Usage: grunt user:delete --email=joe@example.com')
|
||||
process.exit(1)
|
||||
}
|
||||
const settings = require("settings-sharelatex");
|
||||
const mongodb = require("../web/app/src/infrastructure/mongodb");
|
||||
const UserGetter = require("../web/app/src/Features/User/UserGetter");
|
||||
const UserDeleter = require("../web/app/src/Features/User/UserDeleter");
|
||||
return mongodb.waitForDb().then(() => UserGetter.getUser({email}, function(error, user) {
|
||||
const settings = require('settings-sharelatex')
|
||||
const mongodb = require('../web/app/src/infrastructure/mongodb')
|
||||
const UserGetter = require('../web/app/src/Features/User/UserGetter')
|
||||
const UserDeleter = require('../web/app/src/Features/User/UserDeleter')
|
||||
return mongodb.waitForDb().then(() =>
|
||||
UserGetter.getUser({ email }, function (error, user) {
|
||||
if (error != null) {
|
||||
throw error;
|
||||
throw error
|
||||
}
|
||||
if ((user == null)) {
|
||||
console.log(`user ${email} not in database, potentially already deleted`);
|
||||
return done();
|
||||
if (user == null) {
|
||||
console.log(
|
||||
`user ${email} not in database, potentially already deleted`
|
||||
)
|
||||
return done()
|
||||
}
|
||||
return UserDeleter.deleteUser(user._id, function (err) {
|
||||
if (err != null) {
|
||||
throw err;
|
||||
throw err
|
||||
}
|
||||
return done()
|
||||
})
|
||||
})
|
||||
)
|
||||
}
|
||||
)
|
||||
}
|
||||
return done();
|
||||
});
|
||||
}));
|
||||
});
|
||||
};
|
||||
|
|
Loading…
Reference in a new issue