mirror of
https://github.com/overleaf/overleaf.git
synced 2024-10-24 21:12:38 -04:00
1595 lines
43 KiB
JavaScript
1595 lines
43 KiB
JavaScript
|
/* eslint-disable
|
||
|
camelcase,
|
||
|
handle-callback-err,
|
||
|
max-len,
|
||
|
one-var,
|
||
|
standard/no-callback-literal,
|
||
|
*/
|
||
|
// TODO: This file was created by bulk-decaffeinate.
|
||
|
// Fix any style issues and re-enable lint.
|
||
|
/*
|
||
|
* decaffeinate suggestions:
|
||
|
* DS101: Remove unnecessary use of Array.from
|
||
|
* DS102: Remove unnecessary code created because of implicit returns
|
||
|
* DS103: Rewrite code to no longer use __guard__
|
||
|
* DS201: Simplify complex destructure assignments
|
||
|
* DS207: Consider shorter variations of null checks
|
||
|
* Full docs: https://github.com/decaffeinate/decaffeinate/blob/master/docs/suggestions.md
|
||
|
*/
|
||
|
let ProjectEntityUpdateHandler, self
|
||
|
const _ = require('lodash')
|
||
|
const async = require('async')
|
||
|
const logger = require('logger-sharelatex')
|
||
|
const path = require('path')
|
||
|
const { Doc } = require('../../models/Doc')
|
||
|
const DocstoreManager = require('../Docstore/DocstoreManager')
|
||
|
const DocumentUpdaterHandler = require('../../Features/DocumentUpdater/DocumentUpdaterHandler')
|
||
|
const Errors = require('../Errors/Errors')
|
||
|
const { File } = require('../../models/File')
|
||
|
const FileStoreHandler = require('../FileStore/FileStoreHandler')
|
||
|
const LockManager = require('../../infrastructure/LockManager')
|
||
|
const { Project } = require('../../models/Project')
|
||
|
const ProjectEntityHandler = require('./ProjectEntityHandler')
|
||
|
const ProjectGetter = require('./ProjectGetter')
|
||
|
const ProjectLocator = require('./ProjectLocator')
|
||
|
const ProjectUpdateHandler = require('./ProjectUpdateHandler')
|
||
|
const ProjectEntityMongoUpdateHandler = require('./ProjectEntityMongoUpdateHandler')
|
||
|
const SafePath = require('./SafePath')
|
||
|
const TpdsUpdateSender = require('../ThirdPartyDataStore/TpdsUpdateSender')
|
||
|
|
||
|
const LOCK_NAMESPACE = 'sequentialProjectStructureUpdateLock'
|
||
|
|
||
|
const wrapWithLock = function(methodWithoutLock) {
|
||
|
// This lock is used to make sure that the project structure updates are made
|
||
|
// sequentially. In particular the updates must be made in mongo and sent to
|
||
|
// the doc-updater in the same order.
|
||
|
let methodWithLock
|
||
|
if (typeof methodWithoutLock === 'function') {
|
||
|
methodWithLock = function(project_id, ...rest) {
|
||
|
const adjustedLength = Math.max(rest.length, 1),
|
||
|
args = rest.slice(0, adjustedLength - 1),
|
||
|
callback = rest[adjustedLength - 1]
|
||
|
return LockManager.runWithLock(
|
||
|
LOCK_NAMESPACE,
|
||
|
project_id,
|
||
|
cb => methodWithoutLock(project_id, ...Array.from(args), cb),
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
methodWithLock.withoutLock = methodWithoutLock
|
||
|
return methodWithLock
|
||
|
} else {
|
||
|
// handle case with separate setup and locked stages
|
||
|
const wrapWithSetup = methodWithoutLock.beforeLock // a function to set things up before the lock
|
||
|
const mainTask = methodWithoutLock.withLock // function to execute inside the lock
|
||
|
methodWithLock = wrapWithSetup(function(project_id, ...rest) {
|
||
|
const adjustedLength = Math.max(rest.length, 1),
|
||
|
args = rest.slice(0, adjustedLength - 1),
|
||
|
callback = rest[adjustedLength - 1]
|
||
|
return LockManager.runWithLock(
|
||
|
LOCK_NAMESPACE,
|
||
|
project_id,
|
||
|
cb => mainTask(project_id, ...Array.from(args), cb),
|
||
|
callback
|
||
|
)
|
||
|
})
|
||
|
methodWithLock.withoutLock = wrapWithSetup(mainTask)
|
||
|
methodWithLock.beforeLock = methodWithoutLock.beforeLock
|
||
|
methodWithLock.mainTask = methodWithoutLock.withLock
|
||
|
return methodWithLock
|
||
|
}
|
||
|
}
|
||
|
|
||
|
module.exports = ProjectEntityUpdateHandler = self = {
|
||
|
copyFileFromExistingProjectWithProject: wrapWithLock({
|
||
|
beforeLock(next) {
|
||
|
return function(
|
||
|
project_id,
|
||
|
project,
|
||
|
folder_id,
|
||
|
originalProject_id,
|
||
|
origonalFileRef,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error, fileRef, folder_id) {}
|
||
|
}
|
||
|
logger.log(
|
||
|
{ project_id, folder_id, originalProject_id, origonalFileRef },
|
||
|
'copying file in s3 with project'
|
||
|
)
|
||
|
return ProjectEntityMongoUpdateHandler._confirmFolder(
|
||
|
project,
|
||
|
folder_id,
|
||
|
function(folder_id) {
|
||
|
if (origonalFileRef == null) {
|
||
|
logger.err(
|
||
|
{ project_id, folder_id, originalProject_id, origonalFileRef },
|
||
|
'file trying to copy is null'
|
||
|
)
|
||
|
return callback()
|
||
|
}
|
||
|
// convert any invalid characters in original file to '_'
|
||
|
const fileProperties = {
|
||
|
name: SafePath.clean(origonalFileRef.name)
|
||
|
}
|
||
|
if (origonalFileRef.linkedFileData != null) {
|
||
|
fileProperties.linkedFileData = origonalFileRef.linkedFileData
|
||
|
}
|
||
|
if (origonalFileRef.hash != null) {
|
||
|
fileProperties.hash = origonalFileRef.hash
|
||
|
}
|
||
|
const fileRef = new File(fileProperties)
|
||
|
return FileStoreHandler.copyFile(
|
||
|
originalProject_id,
|
||
|
origonalFileRef._id,
|
||
|
project._id,
|
||
|
fileRef._id,
|
||
|
function(err, fileStoreUrl) {
|
||
|
if (err != null) {
|
||
|
logger.err(
|
||
|
{
|
||
|
err,
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
originalProject_id,
|
||
|
origonalFileRef
|
||
|
},
|
||
|
'error coping file in s3'
|
||
|
)
|
||
|
return callback(err)
|
||
|
}
|
||
|
return next(
|
||
|
project_id,
|
||
|
project,
|
||
|
folder_id,
|
||
|
originalProject_id,
|
||
|
origonalFileRef,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
},
|
||
|
withLock(
|
||
|
project_id,
|
||
|
project,
|
||
|
folder_id,
|
||
|
originalProject_id,
|
||
|
origonalFileRef,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error, fileRef, folder_id) {}
|
||
|
}
|
||
|
const projectHistoryId = __guard__(
|
||
|
project.overleaf != null ? project.overleaf.history : undefined,
|
||
|
x => x.id
|
||
|
)
|
||
|
return ProjectEntityMongoUpdateHandler._putElement(
|
||
|
project,
|
||
|
folder_id,
|
||
|
fileRef,
|
||
|
'file',
|
||
|
function(err, result, newProject) {
|
||
|
if (err != null) {
|
||
|
logger.err(
|
||
|
{ err, project_id, folder_id },
|
||
|
'error putting element as part of copy'
|
||
|
)
|
||
|
return callback(err)
|
||
|
}
|
||
|
return TpdsUpdateSender.addFile(
|
||
|
{
|
||
|
project_id,
|
||
|
file_id: fileRef._id,
|
||
|
path: __guard__(
|
||
|
result != null ? result.path : undefined,
|
||
|
x1 => x1.fileSystem
|
||
|
),
|
||
|
rev: fileRef.rev,
|
||
|
project_name: project.name
|
||
|
},
|
||
|
function(err) {
|
||
|
if (err != null) {
|
||
|
logger.err(
|
||
|
{
|
||
|
err,
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
originalProject_id,
|
||
|
origonalFileRef
|
||
|
},
|
||
|
'error sending file to tpds worker'
|
||
|
)
|
||
|
}
|
||
|
const newFiles = [
|
||
|
{
|
||
|
file: fileRef,
|
||
|
path: __guard__(
|
||
|
result != null ? result.path : undefined,
|
||
|
x2 => x2.fileSystem
|
||
|
),
|
||
|
url: fileStoreUrl
|
||
|
}
|
||
|
]
|
||
|
return DocumentUpdaterHandler.updateProjectStructure(
|
||
|
project_id,
|
||
|
projectHistoryId,
|
||
|
userId,
|
||
|
{ newFiles, newProject },
|
||
|
function(error) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return callback(null, fileRef, folder_id)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
}),
|
||
|
|
||
|
updateDocLines(
|
||
|
project_id,
|
||
|
doc_id,
|
||
|
lines,
|
||
|
version,
|
||
|
ranges,
|
||
|
lastUpdatedAt,
|
||
|
lastUpdatedBy,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
return ProjectGetter.getProjectWithoutDocLines(project_id, function(
|
||
|
err,
|
||
|
project
|
||
|
) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
if (project == null) {
|
||
|
return callback(new Errors.NotFoundError('project not found'))
|
||
|
}
|
||
|
logger.log({ project_id, doc_id }, 'updating doc lines')
|
||
|
return ProjectLocator.findElement(
|
||
|
{ project, element_id: doc_id, type: 'docs' },
|
||
|
function(err, doc, path) {
|
||
|
let isDeletedDoc = false
|
||
|
if (err != null) {
|
||
|
if (err instanceof Errors.NotFoundError) {
|
||
|
// We need to be able to update the doclines of deleted docs. This is
|
||
|
// so the doc-updater can flush a doc's content to the doc-store after
|
||
|
// the doc is deleted.
|
||
|
isDeletedDoc = true
|
||
|
doc = _.find(
|
||
|
project.deletedDocs,
|
||
|
doc => doc._id.toString() === doc_id.toString()
|
||
|
)
|
||
|
} else {
|
||
|
return callback(err)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
if (doc == null) {
|
||
|
// Do not allow an update to a doc which has never exist on this project
|
||
|
logger.error(
|
||
|
{ doc_id, project_id, lines },
|
||
|
'doc not found while updating doc lines'
|
||
|
)
|
||
|
return callback(new Errors.NotFoundError('doc not found'))
|
||
|
}
|
||
|
|
||
|
logger.log(
|
||
|
{ project_id, doc_id },
|
||
|
'telling docstore manager to update doc'
|
||
|
)
|
||
|
return DocstoreManager.updateDoc(
|
||
|
project_id,
|
||
|
doc_id,
|
||
|
lines,
|
||
|
version,
|
||
|
ranges,
|
||
|
function(err, modified, rev) {
|
||
|
if (err != null) {
|
||
|
logger.error(
|
||
|
{ err, doc_id, project_id, lines },
|
||
|
'error sending doc to docstore'
|
||
|
)
|
||
|
return callback(err)
|
||
|
}
|
||
|
logger.log(
|
||
|
{ project_id, doc_id, modified },
|
||
|
'finished updating doc lines'
|
||
|
)
|
||
|
// path will only be present if the doc is not deleted
|
||
|
if (modified && !isDeletedDoc) {
|
||
|
// Don't need to block for marking as updated
|
||
|
ProjectUpdateHandler.markAsUpdated(
|
||
|
project_id,
|
||
|
lastUpdatedAt,
|
||
|
lastUpdatedBy
|
||
|
)
|
||
|
return TpdsUpdateSender.addDoc(
|
||
|
{
|
||
|
project_id,
|
||
|
path: path.fileSystem,
|
||
|
doc_id,
|
||
|
project_name: project.name,
|
||
|
rev
|
||
|
},
|
||
|
callback
|
||
|
)
|
||
|
} else {
|
||
|
return callback()
|
||
|
}
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
})
|
||
|
},
|
||
|
|
||
|
setRootDoc(project_id, newRootDocID, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
logger.log({ project_id, rootDocId: newRootDocID }, 'setting root doc')
|
||
|
return Project.update(
|
||
|
{ _id: project_id },
|
||
|
{ rootDoc_id: newRootDocID },
|
||
|
{},
|
||
|
callback
|
||
|
)
|
||
|
},
|
||
|
|
||
|
unsetRootDoc(project_id, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
logger.log({ project_id }, 'removing root doc')
|
||
|
return Project.update(
|
||
|
{ _id: project_id },
|
||
|
{ $unset: { rootDoc_id: true } },
|
||
|
{},
|
||
|
callback
|
||
|
)
|
||
|
},
|
||
|
|
||
|
_addDocAndSendToTpds(project_id, folder_id, doc, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error, result, project) {}
|
||
|
}
|
||
|
return ProjectEntityMongoUpdateHandler.addDoc(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
doc,
|
||
|
function(err, result, project) {
|
||
|
if (err != null) {
|
||
|
logger.err(
|
||
|
{
|
||
|
err,
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
doc_name: doc != null ? doc.name : undefined,
|
||
|
doc_id: doc != null ? doc._id : undefined
|
||
|
},
|
||
|
'error adding file with project'
|
||
|
)
|
||
|
return callback(err)
|
||
|
}
|
||
|
return TpdsUpdateSender.addDoc(
|
||
|
{
|
||
|
project_id,
|
||
|
doc_id: doc != null ? doc._id : undefined,
|
||
|
path: __guard__(
|
||
|
result != null ? result.path : undefined,
|
||
|
x => x.fileSystem
|
||
|
),
|
||
|
project_name: project.name,
|
||
|
rev: 0
|
||
|
},
|
||
|
function(err) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(null, result, project)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
},
|
||
|
|
||
|
addDoc(project_id, folder_id, docName, docLines, userId, callback) {
|
||
|
return self.addDocWithRanges(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
docName,
|
||
|
docLines,
|
||
|
{},
|
||
|
userId,
|
||
|
callback
|
||
|
)
|
||
|
},
|
||
|
|
||
|
addDocWithRanges: wrapWithLock({
|
||
|
beforeLock(next) {
|
||
|
return function(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
docName,
|
||
|
docLines,
|
||
|
ranges,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error, doc, folder_id) {}
|
||
|
}
|
||
|
if (!SafePath.isCleanFilename(docName)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
// Put doc in docstore first, so that if it errors, we don't have a doc_id in the project
|
||
|
// which hasn't been created in docstore.
|
||
|
const doc = new Doc({ name: docName })
|
||
|
return DocstoreManager.updateDoc(
|
||
|
project_id.toString(),
|
||
|
doc._id.toString(),
|
||
|
docLines,
|
||
|
0,
|
||
|
ranges,
|
||
|
function(err, modified, rev) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return next(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
doc,
|
||
|
docName,
|
||
|
docLines,
|
||
|
ranges,
|
||
|
userId,
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
},
|
||
|
withLock(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
doc,
|
||
|
docName,
|
||
|
docLines,
|
||
|
ranges,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error, doc, folder_id) {}
|
||
|
}
|
||
|
return ProjectEntityUpdateHandler._addDocAndSendToTpds(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
doc,
|
||
|
function(err, result, project) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
const docPath = __guard__(
|
||
|
result != null ? result.path : undefined,
|
||
|
x => x.fileSystem
|
||
|
)
|
||
|
const projectHistoryId = __guard__(
|
||
|
project.overleaf != null ? project.overleaf.history : undefined,
|
||
|
x1 => x1.id
|
||
|
)
|
||
|
const newDocs = [
|
||
|
{
|
||
|
doc,
|
||
|
path: docPath,
|
||
|
docLines: docLines.join('\n')
|
||
|
}
|
||
|
]
|
||
|
return DocumentUpdaterHandler.updateProjectStructure(
|
||
|
project_id,
|
||
|
projectHistoryId,
|
||
|
userId,
|
||
|
{ newDocs, newProject: project },
|
||
|
function(error) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return callback(null, doc, folder_id)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
}),
|
||
|
|
||
|
_uploadFile(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error, fileStoreUrl, fileRef) {}
|
||
|
}
|
||
|
if (!SafePath.isCleanFilename(fileName)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
const fileArgs = {
|
||
|
name: fileName,
|
||
|
linkedFileData
|
||
|
}
|
||
|
return FileStoreHandler.uploadFileFromDisk(
|
||
|
project_id,
|
||
|
fileArgs,
|
||
|
fsPath,
|
||
|
function(err, fileStoreUrl, fileRef) {
|
||
|
if (err != null) {
|
||
|
logger.err(
|
||
|
{ err, project_id, folder_id, file_name: fileName, fileRef },
|
||
|
'error uploading image to s3'
|
||
|
)
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(null, fileStoreUrl, fileRef)
|
||
|
}
|
||
|
)
|
||
|
},
|
||
|
|
||
|
_addFileAndSendToTpds(project_id, folder_id, fileRef, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
return ProjectEntityMongoUpdateHandler.addFile(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileRef,
|
||
|
function(err, result, project) {
|
||
|
if (err != null) {
|
||
|
logger.err(
|
||
|
{ err, project_id, folder_id, file_name: fileRef.name, fileRef },
|
||
|
'error adding file with project'
|
||
|
)
|
||
|
return callback(err)
|
||
|
}
|
||
|
return TpdsUpdateSender.addFile(
|
||
|
{
|
||
|
project_id,
|
||
|
file_id: fileRef._id,
|
||
|
path: __guard__(
|
||
|
result != null ? result.path : undefined,
|
||
|
x => x.fileSystem
|
||
|
),
|
||
|
project_name: project.name,
|
||
|
rev: fileRef.rev
|
||
|
},
|
||
|
function(err) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(null, result, project)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
},
|
||
|
|
||
|
addFile: wrapWithLock({
|
||
|
beforeLock(next) {
|
||
|
return function(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (!SafePath.isCleanFilename(fileName)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
return ProjectEntityUpdateHandler._uploadFile(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
function(error, fileStoreUrl, fileRef) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return next(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
},
|
||
|
withLock(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error, fileRef, folder_id) {}
|
||
|
}
|
||
|
return ProjectEntityUpdateHandler._addFileAndSendToTpds(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileRef,
|
||
|
function(err, result, project) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
const projectHistoryId = __guard__(
|
||
|
project.overleaf != null ? project.overleaf.history : undefined,
|
||
|
x => x.id
|
||
|
)
|
||
|
const newFiles = [
|
||
|
{
|
||
|
file: fileRef,
|
||
|
path: __guard__(
|
||
|
result != null ? result.path : undefined,
|
||
|
x1 => x1.fileSystem
|
||
|
),
|
||
|
url: fileStoreUrl
|
||
|
}
|
||
|
]
|
||
|
return DocumentUpdaterHandler.updateProjectStructure(
|
||
|
project_id,
|
||
|
projectHistoryId,
|
||
|
userId,
|
||
|
{ newFiles, newProject: project },
|
||
|
function(error) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return callback(null, fileRef, folder_id)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
}),
|
||
|
|
||
|
replaceFile: wrapWithLock({
|
||
|
beforeLock(next) {
|
||
|
return function(
|
||
|
project_id,
|
||
|
file_id,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
// create a new file
|
||
|
const fileArgs = {
|
||
|
name: 'dummy-upload-filename',
|
||
|
linkedFileData
|
||
|
}
|
||
|
return FileStoreHandler.uploadFileFromDisk(
|
||
|
project_id,
|
||
|
fileArgs,
|
||
|
fsPath,
|
||
|
function(err, fileStoreUrl, fileRef) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return next(
|
||
|
project_id,
|
||
|
file_id,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
},
|
||
|
withLock(
|
||
|
project_id,
|
||
|
file_id,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
newFileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
) {
|
||
|
return ProjectEntityMongoUpdateHandler.replaceFileWithNew(
|
||
|
project_id,
|
||
|
file_id,
|
||
|
newFileRef,
|
||
|
function(err, oldFileRef, project, path, newProject) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
const oldFiles = [
|
||
|
{
|
||
|
file: oldFileRef,
|
||
|
path: path.fileSystem
|
||
|
}
|
||
|
]
|
||
|
const newFiles = [
|
||
|
{
|
||
|
file: newFileRef,
|
||
|
path: path.fileSystem,
|
||
|
url: fileStoreUrl
|
||
|
}
|
||
|
]
|
||
|
const projectHistoryId = __guard__(
|
||
|
project.overleaf != null ? project.overleaf.history : undefined,
|
||
|
x => x.id
|
||
|
)
|
||
|
// Increment the rev for an in-place update (with the same path) so the third-party-datastore
|
||
|
// knows this is a new file.
|
||
|
// Ideally we would get this from ProjectEntityMongoUpdateHandler.replaceFileWithNew
|
||
|
// but it returns the original oldFileRef (after incrementing the rev value in mongo),
|
||
|
// so we add 1 to the rev from that. This isn't atomic and relies on the lock
|
||
|
// but it is acceptable for now.
|
||
|
return TpdsUpdateSender.addFile(
|
||
|
{
|
||
|
project_id: project._id,
|
||
|
file_id: newFileRef._id,
|
||
|
path: path.fileSystem,
|
||
|
rev: oldFileRef.rev + 1,
|
||
|
project_name: project.name
|
||
|
},
|
||
|
function(err) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return DocumentUpdaterHandler.updateProjectStructure(
|
||
|
project_id,
|
||
|
projectHistoryId,
|
||
|
userId,
|
||
|
{ oldFiles, newFiles, newProject },
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
}),
|
||
|
|
||
|
upsertDoc: wrapWithLock(function(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
docName,
|
||
|
docLines,
|
||
|
source,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(err, doc, folder_id, isNewDoc) {}
|
||
|
}
|
||
|
if (!SafePath.isCleanFilename(docName)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
return ProjectLocator.findElement(
|
||
|
{ project_id, element_id: folder_id, type: 'folder' },
|
||
|
function(error, folder) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
if (folder == null) {
|
||
|
return callback(new Error("Couldn't find folder"))
|
||
|
}
|
||
|
let existingDoc = null
|
||
|
for (let doc of Array.from(folder.docs)) {
|
||
|
if (doc.name === docName) {
|
||
|
existingDoc = doc
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
if (existingDoc != null) {
|
||
|
return DocumentUpdaterHandler.setDocument(
|
||
|
project_id,
|
||
|
existingDoc._id,
|
||
|
userId,
|
||
|
docLines,
|
||
|
source,
|
||
|
err => {
|
||
|
logger.log(
|
||
|
{ project_id, doc_id: existingDoc._id },
|
||
|
'notifying users that the document has been updated'
|
||
|
)
|
||
|
return DocumentUpdaterHandler.flushDocToMongo(
|
||
|
project_id,
|
||
|
existingDoc._id,
|
||
|
function(err) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(null, existingDoc, existingDoc == null)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
} else {
|
||
|
return self.addDocWithRanges.withoutLock(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
docName,
|
||
|
docLines,
|
||
|
{},
|
||
|
userId,
|
||
|
function(err, doc) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(null, doc, existingDoc == null)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
}
|
||
|
)
|
||
|
}),
|
||
|
|
||
|
upsertFile: wrapWithLock({
|
||
|
beforeLock(next) {
|
||
|
return function(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (!SafePath.isCleanFilename(fileName)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
// create a new file
|
||
|
const fileArgs = {
|
||
|
name: fileName,
|
||
|
linkedFileData
|
||
|
}
|
||
|
return FileStoreHandler.uploadFileFromDisk(
|
||
|
project_id,
|
||
|
fileArgs,
|
||
|
fsPath,
|
||
|
function(err, fileStoreUrl, fileRef) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return next(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
},
|
||
|
withLock(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
newFileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(err, file, isNewFile, existingFile) {}
|
||
|
}
|
||
|
return ProjectLocator.findElement(
|
||
|
{ project_id, element_id: folder_id, type: 'folder' },
|
||
|
function(error, folder) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
if (folder == null) {
|
||
|
return callback(new Error("Couldn't find folder"))
|
||
|
}
|
||
|
let existingFile = null
|
||
|
for (let fileRef of Array.from(folder.fileRefs)) {
|
||
|
if (fileRef.name === fileName) {
|
||
|
existingFile = fileRef
|
||
|
break
|
||
|
}
|
||
|
}
|
||
|
if (existingFile != null) {
|
||
|
// this calls directly into the replaceFile main task (without the beforeLock part)
|
||
|
return self.replaceFile.mainTask(
|
||
|
project_id,
|
||
|
existingFile._id,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
newFileRef,
|
||
|
fileStoreUrl,
|
||
|
function(err) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(
|
||
|
null,
|
||
|
newFileRef,
|
||
|
existingFile == null,
|
||
|
existingFile
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
} else {
|
||
|
// this calls directly into the addFile main task (without the beforeLock part)
|
||
|
return self.addFile.mainTask(
|
||
|
project_id,
|
||
|
folder_id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
newFileRef,
|
||
|
fileStoreUrl,
|
||
|
function(err) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(
|
||
|
null,
|
||
|
newFileRef,
|
||
|
existingFile == null,
|
||
|
existingFile
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
}),
|
||
|
|
||
|
upsertDocWithPath: wrapWithLock(function(
|
||
|
project_id,
|
||
|
elementPath,
|
||
|
docLines,
|
||
|
source,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (!SafePath.isCleanPath(elementPath)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
const docName = path.basename(elementPath)
|
||
|
const folderPath = path.dirname(elementPath)
|
||
|
return self.mkdirp.withoutLock(project_id, folderPath, function(
|
||
|
err,
|
||
|
newFolders,
|
||
|
folder
|
||
|
) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return self.upsertDoc.withoutLock(
|
||
|
project_id,
|
||
|
folder._id,
|
||
|
docName,
|
||
|
docLines,
|
||
|
source,
|
||
|
userId,
|
||
|
function(err, doc, isNewDoc) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(null, doc, isNewDoc, newFolders, folder)
|
||
|
}
|
||
|
)
|
||
|
})
|
||
|
}),
|
||
|
|
||
|
upsertFileWithPath: wrapWithLock({
|
||
|
beforeLock(next) {
|
||
|
return function(
|
||
|
project_id,
|
||
|
elementPath,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (!SafePath.isCleanPath(elementPath)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
const fileName = path.basename(elementPath)
|
||
|
const folderPath = path.dirname(elementPath)
|
||
|
// create a new file
|
||
|
const fileArgs = {
|
||
|
name: fileName,
|
||
|
linkedFileData
|
||
|
}
|
||
|
return FileStoreHandler.uploadFileFromDisk(
|
||
|
project_id,
|
||
|
fileArgs,
|
||
|
fsPath,
|
||
|
function(err, fileStoreUrl, fileRef) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return next(
|
||
|
project_id,
|
||
|
folderPath,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
},
|
||
|
withLock(
|
||
|
project_id,
|
||
|
folderPath,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
callback
|
||
|
) {
|
||
|
return self.mkdirp.withoutLock(project_id, folderPath, function(
|
||
|
err,
|
||
|
newFolders,
|
||
|
folder
|
||
|
) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
// this calls directly into the upsertFile main task (without the beforeLock part)
|
||
|
return self.upsertFile.mainTask(
|
||
|
project_id,
|
||
|
folder._id,
|
||
|
fileName,
|
||
|
fsPath,
|
||
|
linkedFileData,
|
||
|
userId,
|
||
|
fileRef,
|
||
|
fileStoreUrl,
|
||
|
function(err, newFile, isNewFile, existingFile) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
return callback(
|
||
|
null,
|
||
|
newFile,
|
||
|
isNewFile,
|
||
|
existingFile,
|
||
|
newFolders,
|
||
|
folder
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
})
|
||
|
}
|
||
|
}),
|
||
|
|
||
|
deleteEntity: wrapWithLock(function(
|
||
|
project_id,
|
||
|
entity_id,
|
||
|
entityType,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
logger.log({ entity_id, entityType, project_id }, 'deleting project entity')
|
||
|
if (entityType == null) {
|
||
|
logger.err({ err: 'No entityType set', project_id, entity_id })
|
||
|
return callback('No entityType set')
|
||
|
}
|
||
|
entityType = entityType.toLowerCase()
|
||
|
return ProjectEntityMongoUpdateHandler.deleteEntity(
|
||
|
project_id,
|
||
|
entity_id,
|
||
|
entityType,
|
||
|
function(error, entity, path, projectBeforeDeletion, newProject) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return self._cleanUpEntity(
|
||
|
projectBeforeDeletion,
|
||
|
newProject,
|
||
|
entity,
|
||
|
entityType,
|
||
|
path.fileSystem,
|
||
|
userId,
|
||
|
function(error) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return TpdsUpdateSender.deleteEntity(
|
||
|
{
|
||
|
project_id,
|
||
|
path: path.fileSystem,
|
||
|
project_name: projectBeforeDeletion.name
|
||
|
},
|
||
|
function(error) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return callback(null, entity_id)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}),
|
||
|
|
||
|
deleteEntityWithPath: wrapWithLock((project_id, path, userId, callback) =>
|
||
|
ProjectLocator.findElementByPath({ project_id, path }, function(
|
||
|
err,
|
||
|
element,
|
||
|
type
|
||
|
) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
if (element == null) {
|
||
|
return callback(new Errors.NotFoundError('project not found'))
|
||
|
}
|
||
|
return self.deleteEntity.withoutLock(
|
||
|
project_id,
|
||
|
element._id,
|
||
|
type,
|
||
|
userId,
|
||
|
callback
|
||
|
)
|
||
|
})
|
||
|
),
|
||
|
|
||
|
mkdirp: wrapWithLock(function(project_id, path, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(err, newlyCreatedFolders, lastFolderInPath) {}
|
||
|
}
|
||
|
for (let folder of Array.from(path.split('/'))) {
|
||
|
if (folder.length > 0 && !SafePath.isCleanFilename(folder)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
}
|
||
|
return ProjectEntityMongoUpdateHandler.mkdirp(
|
||
|
project_id,
|
||
|
path,
|
||
|
{ exactCaseMatch: false },
|
||
|
callback
|
||
|
)
|
||
|
}),
|
||
|
|
||
|
mkdirpWithExactCase: wrapWithLock(function(project_id, path, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(err, newlyCreatedFolders, lastFolderInPath) {}
|
||
|
}
|
||
|
for (let folder of Array.from(path.split('/'))) {
|
||
|
if (folder.length > 0 && !SafePath.isCleanFilename(folder)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
}
|
||
|
return ProjectEntityMongoUpdateHandler.mkdirp(
|
||
|
project_id,
|
||
|
path,
|
||
|
{ exactCaseMatch: true },
|
||
|
callback
|
||
|
)
|
||
|
}),
|
||
|
|
||
|
addFolder: wrapWithLock(function(
|
||
|
project_id,
|
||
|
parentFolder_id,
|
||
|
folderName,
|
||
|
callback
|
||
|
) {
|
||
|
if (!SafePath.isCleanFilename(folderName)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
return ProjectEntityMongoUpdateHandler.addFolder(
|
||
|
project_id,
|
||
|
parentFolder_id,
|
||
|
folderName,
|
||
|
callback
|
||
|
)
|
||
|
}),
|
||
|
|
||
|
moveEntity: wrapWithLock(function(
|
||
|
project_id,
|
||
|
entity_id,
|
||
|
destFolderId,
|
||
|
entityType,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
logger.log(
|
||
|
{ entityType, entity_id, project_id, destFolderId },
|
||
|
'moving entity'
|
||
|
)
|
||
|
if (entityType == null) {
|
||
|
logger.err({ err: 'No entityType set', project_id, entity_id })
|
||
|
return callback('No entityType set')
|
||
|
}
|
||
|
entityType = entityType.toLowerCase()
|
||
|
return ProjectEntityMongoUpdateHandler.moveEntity(
|
||
|
project_id,
|
||
|
entity_id,
|
||
|
destFolderId,
|
||
|
entityType,
|
||
|
function(err, project, startPath, endPath, rev, changes) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
const projectHistoryId = __guard__(
|
||
|
project.overleaf != null ? project.overleaf.history : undefined,
|
||
|
x => x.id
|
||
|
)
|
||
|
TpdsUpdateSender.moveEntity({
|
||
|
project_id,
|
||
|
project_name: project.name,
|
||
|
startPath,
|
||
|
endPath,
|
||
|
rev
|
||
|
})
|
||
|
return DocumentUpdaterHandler.updateProjectStructure(
|
||
|
project_id,
|
||
|
projectHistoryId,
|
||
|
userId,
|
||
|
changes,
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}),
|
||
|
|
||
|
renameEntity: wrapWithLock(function(
|
||
|
project_id,
|
||
|
entity_id,
|
||
|
entityType,
|
||
|
newName,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (!SafePath.isCleanFilename(newName)) {
|
||
|
return callback(new Errors.InvalidNameError('invalid element name'))
|
||
|
}
|
||
|
logger.log({ entity_id, project_id }, `renaming ${entityType}`)
|
||
|
if (entityType == null) {
|
||
|
logger.err({ err: 'No entityType set', project_id, entity_id })
|
||
|
return callback('No entityType set')
|
||
|
}
|
||
|
entityType = entityType.toLowerCase()
|
||
|
|
||
|
return ProjectEntityMongoUpdateHandler.renameEntity(
|
||
|
project_id,
|
||
|
entity_id,
|
||
|
entityType,
|
||
|
newName,
|
||
|
function(err, project, startPath, endPath, rev, changes) {
|
||
|
if (err != null) {
|
||
|
return callback(err)
|
||
|
}
|
||
|
const projectHistoryId = __guard__(
|
||
|
project.overleaf != null ? project.overleaf.history : undefined,
|
||
|
x => x.id
|
||
|
)
|
||
|
TpdsUpdateSender.moveEntity({
|
||
|
project_id,
|
||
|
project_name: project.name,
|
||
|
startPath,
|
||
|
endPath,
|
||
|
rev
|
||
|
})
|
||
|
return DocumentUpdaterHandler.updateProjectStructure(
|
||
|
project_id,
|
||
|
projectHistoryId,
|
||
|
userId,
|
||
|
changes,
|
||
|
callback
|
||
|
)
|
||
|
}
|
||
|
)
|
||
|
}),
|
||
|
|
||
|
// This doesn't directly update project structure but we need to take the lock
|
||
|
// to prevent anything else being queued before the resync update
|
||
|
resyncProjectHistory: wrapWithLock((project_id, callback) =>
|
||
|
ProjectGetter.getProject(
|
||
|
project_id,
|
||
|
{ rootFolder: true, overleaf: true },
|
||
|
function(error, project) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
|
||
|
const projectHistoryId = __guard__(
|
||
|
__guard__(
|
||
|
project != null ? project.overleaf : undefined,
|
||
|
x1 => x1.history
|
||
|
),
|
||
|
x => x.id
|
||
|
)
|
||
|
if (projectHistoryId == null) {
|
||
|
error = new Errors.ProjectHistoryDisabledError(
|
||
|
`project history not enabled for ${project_id}`
|
||
|
)
|
||
|
return callback(error)
|
||
|
}
|
||
|
|
||
|
return ProjectEntityHandler.getAllEntitiesFromProject(project, function(
|
||
|
error,
|
||
|
docs,
|
||
|
files
|
||
|
) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
|
||
|
docs = _.map(docs, doc => ({
|
||
|
doc: doc.doc._id,
|
||
|
path: doc.path
|
||
|
}))
|
||
|
|
||
|
files = _.map(files, file => ({
|
||
|
file: file.file._id,
|
||
|
path: file.path,
|
||
|
url: FileStoreHandler._buildUrl(project_id, file.file._id)
|
||
|
}))
|
||
|
|
||
|
return DocumentUpdaterHandler.resyncProjectHistory(
|
||
|
project_id,
|
||
|
projectHistoryId,
|
||
|
docs,
|
||
|
files,
|
||
|
callback
|
||
|
)
|
||
|
})
|
||
|
}
|
||
|
)
|
||
|
),
|
||
|
|
||
|
_cleanUpEntity(
|
||
|
project,
|
||
|
newProject,
|
||
|
entity,
|
||
|
entityType,
|
||
|
path,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
return self._updateProjectStructureWithDeletedEntity(
|
||
|
project,
|
||
|
newProject,
|
||
|
entity,
|
||
|
entityType,
|
||
|
path,
|
||
|
userId,
|
||
|
function(error) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
if (entityType.indexOf('file') !== -1) {
|
||
|
return self._cleanUpFile(project, entity, path, userId, callback)
|
||
|
} else if (entityType.indexOf('doc') !== -1) {
|
||
|
return self._cleanUpDoc(project, entity, path, userId, callback)
|
||
|
} else if (entityType.indexOf('folder') !== -1) {
|
||
|
return self._cleanUpFolder(project, entity, path, userId, callback)
|
||
|
} else {
|
||
|
return callback()
|
||
|
}
|
||
|
}
|
||
|
)
|
||
|
},
|
||
|
|
||
|
// Note: the _cleanUpEntity code and _updateProjectStructureWithDeletedEntity
|
||
|
// methods both need to recursively iterate over the entities in folder.
|
||
|
// These are currently using separate implementations of the recursion. In
|
||
|
// future, these could be simplified using a common project entity iterator.
|
||
|
_updateProjectStructureWithDeletedEntity(
|
||
|
project,
|
||
|
newProject,
|
||
|
entity,
|
||
|
entityType,
|
||
|
entityPath,
|
||
|
userId,
|
||
|
callback
|
||
|
) {
|
||
|
// compute the changes to the project structure
|
||
|
let changes
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
if (entityType.indexOf('file') !== -1) {
|
||
|
changes = { oldFiles: [{ file: entity, path: entityPath }] }
|
||
|
} else if (entityType.indexOf('doc') !== -1) {
|
||
|
changes = { oldDocs: [{ doc: entity, path: entityPath }] }
|
||
|
} else if (entityType.indexOf('folder') !== -1) {
|
||
|
changes = { oldDocs: [], oldFiles: [] }
|
||
|
var _recurseFolder = function(folder, folderPath) {
|
||
|
for (let doc of Array.from(folder.docs)) {
|
||
|
changes.oldDocs.push({ doc, path: path.join(folderPath, doc.name) })
|
||
|
}
|
||
|
for (let file of Array.from(folder.fileRefs)) {
|
||
|
changes.oldFiles.push({
|
||
|
file,
|
||
|
path: path.join(folderPath, file.name)
|
||
|
})
|
||
|
}
|
||
|
return Array.from(folder.folders).map(childFolder =>
|
||
|
_recurseFolder(childFolder, path.join(folderPath, childFolder.name))
|
||
|
)
|
||
|
}
|
||
|
_recurseFolder(entity, entityPath)
|
||
|
}
|
||
|
// now send the project structure changes to the docupdater
|
||
|
changes.newProject = newProject
|
||
|
const project_id = project._id.toString()
|
||
|
const projectHistoryId = __guard__(
|
||
|
project.overleaf != null ? project.overleaf.history : undefined,
|
||
|
x => x.id
|
||
|
)
|
||
|
return DocumentUpdaterHandler.updateProjectStructure(
|
||
|
project_id,
|
||
|
projectHistoryId,
|
||
|
userId,
|
||
|
changes,
|
||
|
callback
|
||
|
)
|
||
|
},
|
||
|
|
||
|
_cleanUpDoc(project, doc, path, userId, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
const project_id = project._id.toString()
|
||
|
const doc_id = doc._id.toString()
|
||
|
const unsetRootDocIfRequired = callback => {
|
||
|
if (
|
||
|
project.rootDoc_id != null &&
|
||
|
project.rootDoc_id.toString() === doc_id
|
||
|
) {
|
||
|
return this.unsetRootDoc(project_id, callback)
|
||
|
} else {
|
||
|
return callback()
|
||
|
}
|
||
|
}
|
||
|
|
||
|
return unsetRootDocIfRequired(function(error) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return ProjectEntityMongoUpdateHandler._insertDeletedDocReference(
|
||
|
project._id,
|
||
|
doc,
|
||
|
function(error) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return DocumentUpdaterHandler.deleteDoc(project_id, doc_id, function(
|
||
|
error
|
||
|
) {
|
||
|
if (error != null) {
|
||
|
return callback(error)
|
||
|
}
|
||
|
return DocstoreManager.deleteDoc(project_id, doc_id, callback)
|
||
|
})
|
||
|
}
|
||
|
)
|
||
|
})
|
||
|
},
|
||
|
|
||
|
_cleanUpFile(project, file, path, userId, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
return ProjectEntityMongoUpdateHandler._insertDeletedFileReference(
|
||
|
project._id,
|
||
|
file,
|
||
|
callback
|
||
|
)
|
||
|
},
|
||
|
|
||
|
_cleanUpFolder(project, folder, folderPath, userId, callback) {
|
||
|
if (callback == null) {
|
||
|
callback = function(error) {}
|
||
|
}
|
||
|
const jobs = []
|
||
|
for (let doc of Array.from(folder.docs)) {
|
||
|
;(function(doc) {
|
||
|
const docPath = path.join(folderPath, doc.name)
|
||
|
return jobs.push(callback =>
|
||
|
self._cleanUpDoc(project, doc, docPath, userId, callback)
|
||
|
)
|
||
|
})(doc)
|
||
|
}
|
||
|
|
||
|
for (let file of Array.from(folder.fileRefs)) {
|
||
|
;(function(file) {
|
||
|
const filePath = path.join(folderPath, file.name)
|
||
|
return jobs.push(callback =>
|
||
|
self._cleanUpFile(project, file, filePath, userId, callback)
|
||
|
)
|
||
|
})(file)
|
||
|
}
|
||
|
|
||
|
for (let childFolder of Array.from(folder.folders)) {
|
||
|
;(function(childFolder) {
|
||
|
folderPath = path.join(folderPath, childFolder.name)
|
||
|
return jobs.push(callback =>
|
||
|
self._cleanUpFolder(
|
||
|
project,
|
||
|
childFolder,
|
||
|
folderPath,
|
||
|
userId,
|
||
|
callback
|
||
|
)
|
||
|
)
|
||
|
})(childFolder)
|
||
|
}
|
||
|
|
||
|
return async.series(jobs, callback)
|
||
|
}
|
||
|
}
|
||
|
|
||
|
function __guard__(value, transform) {
|
||
|
return typeof value !== 'undefined' && value !== null
|
||
|
? transform(value)
|
||
|
: undefined
|
||
|
}
|