mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-21 20:47:08 -05:00
Merge pull request #12205 from overleaf/em-camel-case-docstore
Camel case variables in docstore GitOrigin-RevId: e6c2015cfb63ce125fd32ba8b4f904712b2bb9aa
This commit is contained in:
parent
df3c7e48ab
commit
49f1312b27
8 changed files with 157 additions and 188 deletions
|
@ -48,10 +48,7 @@ async function archiveDoc(projectId, docId) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
logger.debug(
|
logger.debug({ projectId, docId: doc._id }, 'sending doc to persistor')
|
||||||
{ project_id: projectId, doc_id: doc._id },
|
|
||||||
'sending doc to persistor'
|
|
||||||
)
|
|
||||||
const key = `${projectId}/${doc._id}`
|
const key = `${projectId}/${doc._id}`
|
||||||
|
|
||||||
if (doc.lines == null) {
|
if (doc.lines == null) {
|
||||||
|
@ -66,7 +63,7 @@ async function archiveDoc(projectId, docId) {
|
||||||
rangesSize > Settings.max_doc_length
|
rangesSize > Settings.max_doc_length
|
||||||
) {
|
) {
|
||||||
logger.warn(
|
logger.warn(
|
||||||
{ project_id: projectId, doc_id: doc._id, linesSize, rangesSize },
|
{ projectId, docId: doc._id, linesSize, rangesSize },
|
||||||
'large doc found when archiving project'
|
'large doc found when archiving project'
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
/* eslint-disable
|
/* eslint-disable
|
||||||
camelcase,
|
|
||||||
no-dupe-keys,
|
no-dupe-keys,
|
||||||
no-undef,
|
no-undef,
|
||||||
*/
|
*/
|
||||||
|
@ -25,7 +24,7 @@ module.exports = DocManager = {
|
||||||
// collection (which is all that this collection contains). In future, we should
|
// collection (which is all that this collection contains). In future, we should
|
||||||
// migrate this version property to be part of the docs collection, to guarantee
|
// migrate this version property to be part of the docs collection, to guarantee
|
||||||
// consitency between lines and version when writing/reading, and for a simpler schema.
|
// consitency between lines and version when writing/reading, and for a simpler schema.
|
||||||
_getDoc(project_id, doc_id, filter, callback) {
|
_getDoc(projectId, docId, filter, callback) {
|
||||||
if (filter == null) {
|
if (filter == null) {
|
||||||
filter = {}
|
filter = {}
|
||||||
}
|
}
|
||||||
|
@ -36,45 +35,37 @@ module.exports = DocManager = {
|
||||||
return callback(new Error('must include inS3 when getting doc'))
|
return callback(new Error('must include inS3 when getting doc'))
|
||||||
}
|
}
|
||||||
|
|
||||||
return MongoManager.findDoc(
|
return MongoManager.findDoc(projectId, docId, filter, function (err, doc) {
|
||||||
project_id,
|
if (err != null) {
|
||||||
doc_id,
|
return callback(err)
|
||||||
filter,
|
} else if (doc == null) {
|
||||||
function (err, doc) {
|
return callback(
|
||||||
if (err != null) {
|
new Errors.NotFoundError(
|
||||||
return callback(err)
|
`No such doc: ${docId} in project ${projectId}`
|
||||||
} else if (doc == null) {
|
|
||||||
return callback(
|
|
||||||
new Errors.NotFoundError(
|
|
||||||
`No such doc: ${doc_id} in project ${project_id}`
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
} else if (doc != null ? doc.inS3 : undefined) {
|
)
|
||||||
return DocArchive.unarchiveDoc(project_id, doc_id, function (err) {
|
} else if (doc != null ? doc.inS3 : undefined) {
|
||||||
if (err != null) {
|
return DocArchive.unarchiveDoc(projectId, docId, function (err) {
|
||||||
logger.err({ err, project_id, doc_id }, 'error unarchiving doc')
|
if (err != null) {
|
||||||
return callback(err)
|
logger.err({ err, projectId, docId }, 'error unarchiving doc')
|
||||||
|
return callback(err)
|
||||||
|
}
|
||||||
|
return DocManager._getDoc(projectId, docId, filter, callback)
|
||||||
|
})
|
||||||
|
} else {
|
||||||
|
if (filter.version) {
|
||||||
|
return MongoManager.getDocVersion(docId, function (error, version) {
|
||||||
|
if (error != null) {
|
||||||
|
return callback(error)
|
||||||
}
|
}
|
||||||
return DocManager._getDoc(project_id, doc_id, filter, callback)
|
doc.version = version
|
||||||
|
return callback(err, doc)
|
||||||
})
|
})
|
||||||
} else {
|
} else {
|
||||||
if (filter.version) {
|
return callback(err, doc)
|
||||||
return MongoManager.getDocVersion(
|
|
||||||
doc_id,
|
|
||||||
function (error, version) {
|
|
||||||
if (error != null) {
|
|
||||||
return callback(error)
|
|
||||||
}
|
|
||||||
doc.version = version
|
|
||||||
return callback(err, doc)
|
|
||||||
}
|
|
||||||
)
|
|
||||||
} else {
|
|
||||||
return callback(err, doc)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
)
|
})
|
||||||
},
|
},
|
||||||
|
|
||||||
isDocDeleted(projectId, docId, callback) {
|
isDocDeleted(projectId, docId, callback) {
|
||||||
|
@ -99,13 +90,13 @@ module.exports = DocManager = {
|
||||||
)
|
)
|
||||||
},
|
},
|
||||||
|
|
||||||
getFullDoc(project_id, doc_id, callback) {
|
getFullDoc(projectId, docId, callback) {
|
||||||
if (callback == null) {
|
if (callback == null) {
|
||||||
callback = function () {}
|
callback = function () {}
|
||||||
}
|
}
|
||||||
return DocManager._getDoc(
|
return DocManager._getDoc(
|
||||||
project_id,
|
projectId,
|
||||||
doc_id,
|
docId,
|
||||||
{
|
{
|
||||||
lines: true,
|
lines: true,
|
||||||
rev: true,
|
rev: true,
|
||||||
|
@ -124,10 +115,10 @@ module.exports = DocManager = {
|
||||||
},
|
},
|
||||||
|
|
||||||
// returns the doc without any version information
|
// returns the doc without any version information
|
||||||
_peekRawDoc(project_id, doc_id, callback) {
|
_peekRawDoc(projectId, docId, callback) {
|
||||||
MongoManager.findDoc(
|
MongoManager.findDoc(
|
||||||
project_id,
|
projectId,
|
||||||
doc_id,
|
docId,
|
||||||
{
|
{
|
||||||
lines: true,
|
lines: true,
|
||||||
rev: true,
|
rev: true,
|
||||||
|
@ -141,7 +132,7 @@ module.exports = DocManager = {
|
||||||
if (doc == null) {
|
if (doc == null) {
|
||||||
return callback(
|
return callback(
|
||||||
new Errors.NotFoundError(
|
new Errors.NotFoundError(
|
||||||
`No such doc: ${doc_id} in project ${project_id}`
|
`No such doc: ${docId} in project ${projectId}`
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
@ -149,10 +140,10 @@ module.exports = DocManager = {
|
||||||
return callback(null, doc)
|
return callback(null, doc)
|
||||||
}
|
}
|
||||||
// skip the unarchiving to mongo when getting a doc
|
// skip the unarchiving to mongo when getting a doc
|
||||||
DocArchive.getDoc(project_id, doc_id, function (err, archivedDoc) {
|
DocArchive.getDoc(projectId, docId, function (err, archivedDoc) {
|
||||||
if (err != null) {
|
if (err != null) {
|
||||||
logger.err(
|
logger.err(
|
||||||
{ err, project_id, doc_id },
|
{ err, projectId, docId },
|
||||||
'error getting doc from archive'
|
'error getting doc from archive'
|
||||||
)
|
)
|
||||||
return callback(err)
|
return callback(err)
|
||||||
|
@ -166,8 +157,8 @@ module.exports = DocManager = {
|
||||||
|
|
||||||
// get the doc from mongo if possible, or from the persistent store otherwise,
|
// get the doc from mongo if possible, or from the persistent store otherwise,
|
||||||
// without unarchiving it (avoids unnecessary writes to mongo)
|
// without unarchiving it (avoids unnecessary writes to mongo)
|
||||||
peekDoc(project_id, doc_id, callback) {
|
peekDoc(projectId, docId, callback) {
|
||||||
DocManager._peekRawDoc(project_id, doc_id, (err, doc) => {
|
DocManager._peekRawDoc(projectId, docId, (err, doc) => {
|
||||||
if (err) {
|
if (err) {
|
||||||
return callback(err)
|
return callback(err)
|
||||||
}
|
}
|
||||||
|
@ -187,13 +178,13 @@ module.exports = DocManager = {
|
||||||
})
|
})
|
||||||
},
|
},
|
||||||
|
|
||||||
getDocLines(project_id, doc_id, callback) {
|
getDocLines(projectId, docId, callback) {
|
||||||
if (callback == null) {
|
if (callback == null) {
|
||||||
callback = function () {}
|
callback = function () {}
|
||||||
}
|
}
|
||||||
return DocManager._getDoc(
|
return DocManager._getDoc(
|
||||||
project_id,
|
projectId,
|
||||||
doc_id,
|
docId,
|
||||||
{ lines: true, inS3: true },
|
{ lines: true, inS3: true },
|
||||||
function (err, doc) {
|
function (err, doc) {
|
||||||
if (err != null) {
|
if (err != null) {
|
||||||
|
@ -204,20 +195,20 @@ module.exports = DocManager = {
|
||||||
)
|
)
|
||||||
},
|
},
|
||||||
|
|
||||||
getAllDeletedDocs(project_id, filter, callback) {
|
getAllDeletedDocs(projectId, filter, callback) {
|
||||||
MongoManager.getProjectsDeletedDocs(project_id, filter, callback)
|
MongoManager.getProjectsDeletedDocs(projectId, filter, callback)
|
||||||
},
|
},
|
||||||
|
|
||||||
getAllNonDeletedDocs(project_id, filter, callback) {
|
getAllNonDeletedDocs(projectId, filter, callback) {
|
||||||
if (callback == null) {
|
if (callback == null) {
|
||||||
callback = function () {}
|
callback = function () {}
|
||||||
}
|
}
|
||||||
return DocArchive.unArchiveAllDocs(project_id, function (error) {
|
return DocArchive.unArchiveAllDocs(projectId, function (error) {
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
return callback(error)
|
return callback(error)
|
||||||
}
|
}
|
||||||
return MongoManager.getProjectsDocs(
|
return MongoManager.getProjectsDocs(
|
||||||
project_id,
|
projectId,
|
||||||
{ include_deleted: false },
|
{ include_deleted: false },
|
||||||
filter,
|
filter,
|
||||||
function (error, docs) {
|
function (error, docs) {
|
||||||
|
@ -225,7 +216,7 @@ module.exports = DocManager = {
|
||||||
return callback(error)
|
return callback(error)
|
||||||
} else if (docs == null) {
|
} else if (docs == null) {
|
||||||
return callback(
|
return callback(
|
||||||
new Errors.NotFoundError(`No docs for project ${project_id}`)
|
new Errors.NotFoundError(`No docs for project ${projectId}`)
|
||||||
)
|
)
|
||||||
} else {
|
} else {
|
||||||
return callback(null, docs)
|
return callback(null, docs)
|
||||||
|
@ -235,7 +226,7 @@ module.exports = DocManager = {
|
||||||
})
|
})
|
||||||
},
|
},
|
||||||
|
|
||||||
updateDoc(project_id, doc_id, lines, version, ranges, callback) {
|
updateDoc(projectId, docId, lines, version, ranges, callback) {
|
||||||
if (callback == null) {
|
if (callback == null) {
|
||||||
callback = function () {}
|
callback = function () {}
|
||||||
}
|
}
|
||||||
|
@ -244,8 +235,8 @@ module.exports = DocManager = {
|
||||||
}
|
}
|
||||||
|
|
||||||
return DocManager._getDoc(
|
return DocManager._getDoc(
|
||||||
project_id,
|
projectId,
|
||||||
doc_id,
|
docId,
|
||||||
{
|
{
|
||||||
version: true,
|
version: true,
|
||||||
rev: true,
|
rev: true,
|
||||||
|
@ -258,7 +249,7 @@ module.exports = DocManager = {
|
||||||
let updateLines, updateRanges, updateVersion
|
let updateLines, updateRanges, updateVersion
|
||||||
if (err != null && !(err instanceof Errors.NotFoundError)) {
|
if (err != null && !(err instanceof Errors.NotFoundError)) {
|
||||||
logger.err(
|
logger.err(
|
||||||
{ project_id, doc_id, err },
|
{ projectId, docId, err },
|
||||||
'error getting document for update'
|
'error getting document for update'
|
||||||
)
|
)
|
||||||
return callback(err)
|
return callback(err)
|
||||||
|
@ -289,22 +280,19 @@ module.exports = DocManager = {
|
||||||
if (updateRanges) {
|
if (updateRanges) {
|
||||||
update.ranges = ranges
|
update.ranges = ranges
|
||||||
}
|
}
|
||||||
logger.debug(
|
logger.debug({ projectId, docId }, 'updating doc lines and ranges')
|
||||||
{ project_id, doc_id },
|
|
||||||
'updating doc lines and ranges'
|
|
||||||
)
|
|
||||||
|
|
||||||
modified = true
|
modified = true
|
||||||
rev += 1 // rev will be incremented in mongo by MongoManager.upsertIntoDocCollection
|
rev += 1 // rev will be incremented in mongo by MongoManager.upsertIntoDocCollection
|
||||||
return MongoManager.upsertIntoDocCollection(
|
return MongoManager.upsertIntoDocCollection(
|
||||||
project_id,
|
projectId,
|
||||||
doc_id,
|
docId,
|
||||||
update,
|
update,
|
||||||
cb
|
cb
|
||||||
)
|
)
|
||||||
} else {
|
} else {
|
||||||
logger.debug(
|
logger.debug(
|
||||||
{ project_id, doc_id },
|
{ projectId, docId },
|
||||||
'doc lines have not changed - not updating'
|
'doc lines have not changed - not updating'
|
||||||
)
|
)
|
||||||
return cb()
|
return cb()
|
||||||
|
@ -315,18 +303,18 @@ module.exports = DocManager = {
|
||||||
if (updateVersion) {
|
if (updateVersion) {
|
||||||
logger.debug(
|
logger.debug(
|
||||||
{
|
{
|
||||||
project_id,
|
projectId,
|
||||||
doc_id,
|
docId,
|
||||||
oldVersion: doc != null ? doc.version : undefined,
|
oldVersion: doc != null ? doc.version : undefined,
|
||||||
newVersion: version,
|
newVersion: version,
|
||||||
},
|
},
|
||||||
'updating doc version'
|
'updating doc version'
|
||||||
)
|
)
|
||||||
modified = true
|
modified = true
|
||||||
return MongoManager.setDocVersion(doc_id, version, cb)
|
return MongoManager.setDocVersion(docId, version, cb)
|
||||||
} else {
|
} else {
|
||||||
logger.debug(
|
logger.debug(
|
||||||
{ project_id, doc_id, version },
|
{ projectId, docId, version },
|
||||||
'doc version has not changed - not updating'
|
'doc version has not changed - not updating'
|
||||||
)
|
)
|
||||||
return cb()
|
return cb()
|
||||||
|
@ -348,33 +336,33 @@ module.exports = DocManager = {
|
||||||
)
|
)
|
||||||
},
|
},
|
||||||
|
|
||||||
patchDoc(project_id, doc_id, meta, callback) {
|
patchDoc(projectId, docId, meta, callback) {
|
||||||
const projection = { _id: 1, deleted: true }
|
const projection = { _id: 1, deleted: true }
|
||||||
MongoManager.findDoc(project_id, doc_id, projection, (error, doc) => {
|
MongoManager.findDoc(projectId, docId, projection, (error, doc) => {
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
return callback(error)
|
return callback(error)
|
||||||
}
|
}
|
||||||
if (!doc) {
|
if (!doc) {
|
||||||
return callback(
|
return callback(
|
||||||
new Errors.NotFoundError(
|
new Errors.NotFoundError(
|
||||||
`No such project/doc to delete: ${project_id}/${doc_id}`
|
`No such project/doc to delete: ${projectId}/${docId}`
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (meta.deleted && Settings.docstore.archiveOnSoftDelete) {
|
if (meta.deleted && Settings.docstore.archiveOnSoftDelete) {
|
||||||
// The user will not read this doc anytime soon. Flush it out of mongo.
|
// The user will not read this doc anytime soon. Flush it out of mongo.
|
||||||
DocArchive.archiveDoc(project_id, doc_id, err => {
|
DocArchive.archiveDoc(projectId, docId, err => {
|
||||||
if (err) {
|
if (err) {
|
||||||
logger.warn(
|
logger.warn(
|
||||||
{ project_id, doc_id, err },
|
{ projectId, docId, err },
|
||||||
'archiving a single doc in the background failed'
|
'archiving a single doc in the background failed'
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
MongoManager.patchDoc(project_id, doc_id, meta, callback)
|
MongoManager.patchDoc(projectId, docId, meta, callback)
|
||||||
})
|
})
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,6 +1,3 @@
|
||||||
/* eslint-disable
|
|
||||||
camelcase,
|
|
||||||
*/
|
|
||||||
// TODO: This file was created by bulk-decaffeinate.
|
// TODO: This file was created by bulk-decaffeinate.
|
||||||
// Fix any style issues and re-enable lint.
|
// Fix any style issues and re-enable lint.
|
||||||
/*
|
/*
|
||||||
|
@ -20,9 +17,9 @@ const logger = require('@overleaf/logger')
|
||||||
|
|
||||||
module.exports = {
|
module.exports = {
|
||||||
check(callback) {
|
check(callback) {
|
||||||
const doc_id = ObjectId()
|
const docId = ObjectId()
|
||||||
const project_id = ObjectId(settings.docstore.healthCheck.project_id)
|
const projectId = ObjectId(settings.docstore.healthCheck.project_id)
|
||||||
const url = `http://localhost:${port}/project/${project_id}/doc/${doc_id}`
|
const url = `http://localhost:${port}/project/${projectId}/doc/${docId}`
|
||||||
const lines = [
|
const lines = [
|
||||||
'smoke test - delete me',
|
'smoke test - delete me',
|
||||||
`${crypto.randomBytes(32).toString('hex')}`,
|
`${crypto.randomBytes(32).toString('hex')}`,
|
||||||
|
@ -31,7 +28,7 @@ module.exports = {
|
||||||
url,
|
url,
|
||||||
timeout: 3000,
|
timeout: 3000,
|
||||||
})
|
})
|
||||||
logger.debug({ lines, url, doc_id, project_id }, 'running health check')
|
logger.debug({ lines, url, docId, projectId }, 'running health check')
|
||||||
const jobs = [
|
const jobs = [
|
||||||
function (cb) {
|
function (cb) {
|
||||||
const opts = getOpts()
|
const opts = getOpts()
|
||||||
|
@ -51,7 +48,7 @@ module.exports = {
|
||||||
return cb(new Error(`status code not 200, its ${res.statusCode}`))
|
return cb(new Error(`status code not 200, its ${res.statusCode}`))
|
||||||
} else if (
|
} else if (
|
||||||
_.isEqual(body != null ? body.lines : undefined, lines) &&
|
_.isEqual(body != null ? body.lines : undefined, lines) &&
|
||||||
(body != null ? body._id : undefined) === doc_id.toString()
|
(body != null ? body._id : undefined) === docId.toString()
|
||||||
) {
|
) {
|
||||||
return cb()
|
return cb()
|
||||||
} else {
|
} else {
|
||||||
|
@ -63,8 +60,8 @@ module.exports = {
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
},
|
},
|
||||||
cb => db.docs.deleteOne({ _id: doc_id, project_id }, cb),
|
cb => db.docs.deleteOne({ _id: docId, project_id: projectId }, cb),
|
||||||
cb => db.docOps.deleteOne({ doc_id }, cb),
|
cb => db.docOps.deleteOne({ doc_id: docId }, cb),
|
||||||
]
|
]
|
||||||
return async.series(jobs, callback)
|
return async.series(jobs, callback)
|
||||||
},
|
},
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
/* eslint-disable
|
/* eslint-disable
|
||||||
camelcase,
|
|
||||||
no-return-assign,
|
no-return-assign,
|
||||||
*/
|
*/
|
||||||
// TODO: This file was created by bulk-decaffeinate.
|
// TODO: This file was created by bulk-decaffeinate.
|
||||||
|
@ -16,19 +15,19 @@ const _ = require('lodash')
|
||||||
const { ObjectId } = require('./mongodb')
|
const { ObjectId } = require('./mongodb')
|
||||||
|
|
||||||
module.exports = RangeManager = {
|
module.exports = RangeManager = {
|
||||||
shouldUpdateRanges(doc_ranges, incoming_ranges) {
|
shouldUpdateRanges(docRanges, incomingRanges) {
|
||||||
if (incoming_ranges == null) {
|
if (incomingRanges == null) {
|
||||||
throw new Error('expected incoming_ranges')
|
throw new Error('expected incoming_ranges')
|
||||||
}
|
}
|
||||||
|
|
||||||
// If the ranges are empty, we don't store them in the DB, so set
|
// If the ranges are empty, we don't store them in the DB, so set
|
||||||
// doc_ranges to an empty object as default, since this is was the
|
// doc_ranges to an empty object as default, since this is was the
|
||||||
// incoming_ranges will be for an empty range set.
|
// incoming_ranges will be for an empty range set.
|
||||||
if (doc_ranges == null) {
|
if (docRanges == null) {
|
||||||
doc_ranges = {}
|
docRanges = {}
|
||||||
}
|
}
|
||||||
|
|
||||||
return !_.isEqual(doc_ranges, incoming_ranges)
|
return !_.isEqual(docRanges, incomingRanges)
|
||||||
},
|
},
|
||||||
|
|
||||||
jsonRangesToMongo(ranges) {
|
jsonRangesToMongo(ranges) {
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
/* eslint-disable
|
/* eslint-disable
|
||||||
camelcase,
|
|
||||||
no-unused-vars,
|
no-unused-vars,
|
||||||
*/
|
*/
|
||||||
// TODO: This file was created by bulk-decaffeinate.
|
// TODO: This file was created by bulk-decaffeinate.
|
||||||
|
@ -113,10 +112,10 @@ describe('Archiving', function () {
|
||||||
return DocstoreClient.getS3Doc(
|
return DocstoreClient.getS3Doc(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
doc._id,
|
doc._id,
|
||||||
(error, s3_doc) => {
|
(error, s3Doc) => {
|
||||||
if (error) return callback(error)
|
if (error) return callback(error)
|
||||||
s3_doc.lines.should.deep.equal(doc.lines)
|
s3Doc.lines.should.deep.equal(doc.lines)
|
||||||
s3_doc.ranges.should.deep.equal(doc.ranges)
|
s3Doc.ranges.should.deep.equal(doc.ranges)
|
||||||
callback()
|
callback()
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -130,8 +129,8 @@ describe('Archiving', function () {
|
||||||
before(function (done) {
|
before(function (done) {
|
||||||
return DocstoreClient.getAllDocs(
|
return DocstoreClient.getAllDocs(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
(error, res, fetched_docs) => {
|
(error, res, fetchedDocs) => {
|
||||||
this.fetched_docs = fetched_docs
|
this.fetched_docs = fetchedDocs
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
|
@ -231,12 +230,12 @@ describe('Archiving', function () {
|
||||||
return DocstoreClient.getS3Doc(
|
return DocstoreClient.getS3Doc(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
this.doc._id,
|
this.doc._id,
|
||||||
(error, s3_doc) => {
|
(error, s3Doc) => {
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
s3_doc.lines.should.deep.equal(this.doc.lines)
|
s3Doc.lines.should.deep.equal(this.doc.lines)
|
||||||
s3_doc.ranges.should.deep.equal(this.doc.ranges)
|
s3Doc.ranges.should.deep.equal(this.doc.ranges)
|
||||||
return done()
|
return done()
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -246,8 +245,8 @@ describe('Archiving', function () {
|
||||||
beforeEach(function (done) {
|
beforeEach(function (done) {
|
||||||
return DocstoreClient.getAllDocs(
|
return DocstoreClient.getAllDocs(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
(error, res, fetched_docs) => {
|
(error, res, fetchedDocs) => {
|
||||||
this.fetched_docs = fetched_docs
|
this.fetched_docs = fetchedDocs
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
|
@ -291,8 +290,8 @@ describe('Archiving', function () {
|
||||||
beforeEach(function (done) {
|
beforeEach(function (done) {
|
||||||
DocstoreClient.getAllDocs(
|
DocstoreClient.getAllDocs(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
(error, res, fetched_docs) => {
|
(error, res, fetchedDocs) => {
|
||||||
this.fetched_docs = fetched_docs
|
this.fetched_docs = fetchedDocs
|
||||||
if (error) {
|
if (error) {
|
||||||
return done(error)
|
return done(error)
|
||||||
}
|
}
|
||||||
|
@ -375,18 +374,14 @@ describe('Archiving', function () {
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should set the doc in s3 correctly', function (done) {
|
it('should set the doc in s3 correctly', function (done) {
|
||||||
DocstoreClient.getS3Doc(
|
DocstoreClient.getS3Doc(this.project_id, this.doc._id, (error, s3Doc) => {
|
||||||
this.project_id,
|
if (error) {
|
||||||
this.doc._id,
|
return done(error)
|
||||||
(error, s3_doc) => {
|
|
||||||
if (error) {
|
|
||||||
return done(error)
|
|
||||||
}
|
|
||||||
s3_doc.lines.should.deep.equal(this.doc.lines)
|
|
||||||
s3_doc.ranges.should.deep.equal(this.doc.ranges)
|
|
||||||
done()
|
|
||||||
}
|
}
|
||||||
)
|
s3Doc.lines.should.deep.equal(this.doc.lines)
|
||||||
|
s3Doc.ranges.should.deep.equal(this.doc.ranges)
|
||||||
|
done()
|
||||||
|
})
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -395,12 +390,12 @@ describe('Archiving', function () {
|
||||||
this.project_id = ObjectId()
|
this.project_id = ObjectId()
|
||||||
this.timeout(1000 * 30)
|
this.timeout(1000 * 30)
|
||||||
const quarterMegInBytes = 250000
|
const quarterMegInBytes = 250000
|
||||||
const big_line = require('crypto')
|
const bigLine = require('crypto')
|
||||||
.randomBytes(quarterMegInBytes)
|
.randomBytes(quarterMegInBytes)
|
||||||
.toString('hex')
|
.toString('hex')
|
||||||
this.doc = {
|
this.doc = {
|
||||||
_id: ObjectId(),
|
_id: ObjectId(),
|
||||||
lines: [big_line, big_line, big_line, big_line],
|
lines: [bigLine, bigLine, bigLine, bigLine],
|
||||||
ranges: {},
|
ranges: {},
|
||||||
version: 2,
|
version: 2,
|
||||||
}
|
}
|
||||||
|
@ -446,12 +441,12 @@ describe('Archiving', function () {
|
||||||
return DocstoreClient.getS3Doc(
|
return DocstoreClient.getS3Doc(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
this.doc._id,
|
this.doc._id,
|
||||||
(error, s3_doc) => {
|
(error, s3Doc) => {
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
s3_doc.lines.should.deep.equal(this.doc.lines)
|
s3Doc.lines.should.deep.equal(this.doc.lines)
|
||||||
s3_doc.ranges.should.deep.equal(this.doc.ranges)
|
s3Doc.ranges.should.deep.equal(this.doc.ranges)
|
||||||
return done()
|
return done()
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -461,8 +456,8 @@ describe('Archiving', function () {
|
||||||
before(function (done) {
|
before(function (done) {
|
||||||
return DocstoreClient.getAllDocs(
|
return DocstoreClient.getAllDocs(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
(error, res, fetched_docs) => {
|
(error, res, fetchedDocs) => {
|
||||||
this.fetched_docs = fetched_docs
|
this.fetched_docs = fetchedDocs
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
|
@ -920,12 +915,12 @@ describe('Archiving', function () {
|
||||||
return DocstoreClient.getS3Doc(
|
return DocstoreClient.getS3Doc(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
this.doc._id,
|
this.doc._id,
|
||||||
(error, s3_doc) => {
|
(error, s3Doc) => {
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
s3_doc.lines.should.deep.equal(this.doc.lines)
|
s3Doc.lines.should.deep.equal(this.doc.lines)
|
||||||
s3_doc.ranges.should.deep.equal(this.doc.ranges)
|
s3Doc.ranges.should.deep.equal(this.doc.ranges)
|
||||||
return done()
|
return done()
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -935,8 +930,8 @@ describe('Archiving', function () {
|
||||||
before(function (done) {
|
before(function (done) {
|
||||||
return DocstoreClient.getAllDocs(
|
return DocstoreClient.getAllDocs(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
(error, res, fetched_docs) => {
|
(error, res, fetchedDocs) => {
|
||||||
this.fetched_docs = fetched_docs
|
this.fetched_docs = fetchedDocs
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
|
@ -1039,13 +1034,13 @@ describe('Archiving', function () {
|
||||||
return DocstoreClient.getS3Doc(
|
return DocstoreClient.getS3Doc(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
this.doc._id,
|
this.doc._id,
|
||||||
(error, s3_doc) => {
|
(error, s3Doc) => {
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
s3_doc.lines.should.deep.equal(this.doc.lines)
|
s3Doc.lines.should.deep.equal(this.doc.lines)
|
||||||
const ranges = JSON.parse(JSON.stringify(this.doc.ranges)) // ObjectId -> String
|
const ranges = JSON.parse(JSON.stringify(this.doc.ranges)) // ObjectId -> String
|
||||||
s3_doc.ranges.should.deep.equal(ranges)
|
s3Doc.ranges.should.deep.equal(ranges)
|
||||||
return done()
|
return done()
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -1055,8 +1050,8 @@ describe('Archiving', function () {
|
||||||
before(function (done) {
|
before(function (done) {
|
||||||
return DocstoreClient.getAllDocs(
|
return DocstoreClient.getAllDocs(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
(error, res, fetched_docs) => {
|
(error, res, fetchedDocs) => {
|
||||||
this.fetched_docs = fetched_docs
|
this.fetched_docs = fetchedDocs
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
|
@ -1136,12 +1131,12 @@ describe('Archiving', function () {
|
||||||
return DocstoreClient.getS3Doc(
|
return DocstoreClient.getS3Doc(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
this.doc._id,
|
this.doc._id,
|
||||||
(error, s3_doc) => {
|
(error, s3Doc) => {
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
s3_doc.lines.should.deep.equal(this.doc.lines)
|
s3Doc.lines.should.deep.equal(this.doc.lines)
|
||||||
s3_doc.ranges.should.deep.equal(this.doc.ranges)
|
s3Doc.ranges.should.deep.equal(this.doc.ranges)
|
||||||
return done()
|
return done()
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -1151,8 +1146,8 @@ describe('Archiving', function () {
|
||||||
before(function (done) {
|
before(function (done) {
|
||||||
return DocstoreClient.getAllDocs(
|
return DocstoreClient.getAllDocs(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
(error, res, fetched_docs) => {
|
(error, res, fetchedDocs) => {
|
||||||
this.fetched_docs = fetched_docs
|
this.fetched_docs = fetchedDocs
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
|
@ -1202,8 +1197,8 @@ describe('Archiving', function () {
|
||||||
}
|
}
|
||||||
DocstoreClient.getAllDocs(
|
DocstoreClient.getAllDocs(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
(error, res, fetched_docs) => {
|
(error, res, fetchedDocs) => {
|
||||||
this.fetched_docs = fetched_docs
|
this.fetched_docs = fetchedDocs
|
||||||
if (error != null) {
|
if (error != null) {
|
||||||
throw error
|
throw error
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
/* eslint-disable
|
/* eslint-disable
|
||||||
camelcase,
|
|
||||||
no-unused-vars,
|
no-unused-vars,
|
||||||
*/
|
*/
|
||||||
// TODO: This file was created by bulk-decaffeinate.
|
// TODO: This file was created by bulk-decaffeinate.
|
||||||
|
@ -70,10 +69,10 @@ describe('Getting a doc', function () {
|
||||||
|
|
||||||
describe('when the doc does not exist', function () {
|
describe('when the doc does not exist', function () {
|
||||||
return it('should return a 404', function (done) {
|
return it('should return a 404', function (done) {
|
||||||
const missing_doc_id = ObjectId()
|
const missingDocId = ObjectId()
|
||||||
return DocstoreClient.getDoc(
|
return DocstoreClient.getDoc(
|
||||||
this.project_id,
|
this.project_id,
|
||||||
missing_doc_id,
|
missingDocId,
|
||||||
{},
|
{},
|
||||||
(error, res, doc) => {
|
(error, res, doc) => {
|
||||||
if (error) return done(error)
|
if (error) return done(error)
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
/* eslint-disable
|
/* eslint-disable
|
||||||
camelcase,
|
|
||||||
no-dupe-keys,
|
no-dupe-keys,
|
||||||
no-return-assign,
|
no-return-assign,
|
||||||
no-unused-vars,
|
no-unused-vars,
|
||||||
|
@ -254,11 +253,7 @@ describe('DocManager', function () {
|
||||||
inS3: true,
|
inS3: true,
|
||||||
}
|
}
|
||||||
this.MongoManager.findDoc.yields(null, this.doc)
|
this.MongoManager.findDoc.yields(null, this.doc)
|
||||||
this.DocArchiveManager.unarchiveDoc = (
|
this.DocArchiveManager.unarchiveDoc = (projectId, docId, callback) => {
|
||||||
project_id,
|
|
||||||
doc_id,
|
|
||||||
callback
|
|
||||||
) => {
|
|
||||||
this.doc.inS3 = false
|
this.doc.inS3 = false
|
||||||
return callback()
|
return callback()
|
||||||
}
|
}
|
||||||
|
@ -497,8 +492,8 @@ describe('DocManager', function () {
|
||||||
it('should log a warning', function () {
|
it('should log a warning', function () {
|
||||||
expect(this.logger.warn).to.have.been.calledWith(
|
expect(this.logger.warn).to.have.been.calledWith(
|
||||||
sinon.match({
|
sinon.match({
|
||||||
project_id: this.project_id,
|
projectId: this.project_id,
|
||||||
doc_id: this.doc_id,
|
docId: this.doc_id,
|
||||||
err: this.err,
|
err: this.err,
|
||||||
}),
|
}),
|
||||||
'archiving a single doc in the background failed'
|
'archiving a single doc in the background failed'
|
||||||
|
|
|
@ -1,5 +1,4 @@
|
||||||
/* eslint-disable
|
/* eslint-disable
|
||||||
camelcase,
|
|
||||||
no-return-assign,
|
no-return-assign,
|
||||||
no-unused-vars,
|
no-unused-vars,
|
||||||
*/
|
*/
|
||||||
|
@ -32,110 +31,110 @@ describe('RangeManager', function () {
|
||||||
|
|
||||||
describe('jsonRangesToMongo', function () {
|
describe('jsonRangesToMongo', function () {
|
||||||
it('should convert ObjectIds and dates to proper objects', function () {
|
it('should convert ObjectIds and dates to proper objects', function () {
|
||||||
const change_id = ObjectId().toString()
|
const changeId = ObjectId().toString()
|
||||||
const comment_id = ObjectId().toString()
|
const commentId = ObjectId().toString()
|
||||||
const user_id = ObjectId().toString()
|
const userId = ObjectId().toString()
|
||||||
const thread_id = ObjectId().toString()
|
const threadId = ObjectId().toString()
|
||||||
const ts = new Date().toJSON()
|
const ts = new Date().toJSON()
|
||||||
return this.RangeManager.jsonRangesToMongo({
|
return this.RangeManager.jsonRangesToMongo({
|
||||||
changes: [
|
changes: [
|
||||||
{
|
{
|
||||||
id: change_id,
|
id: changeId,
|
||||||
op: { i: 'foo', p: 3 },
|
op: { i: 'foo', p: 3 },
|
||||||
metadata: {
|
metadata: {
|
||||||
user_id,
|
user_id: userId,
|
||||||
ts,
|
ts,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
comments: [
|
comments: [
|
||||||
{
|
{
|
||||||
id: comment_id,
|
id: commentId,
|
||||||
op: { c: 'foo', p: 3, t: thread_id },
|
op: { c: 'foo', p: 3, t: threadId },
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
}).should.deep.equal({
|
}).should.deep.equal({
|
||||||
changes: [
|
changes: [
|
||||||
{
|
{
|
||||||
id: ObjectId(change_id),
|
id: ObjectId(changeId),
|
||||||
op: { i: 'foo', p: 3 },
|
op: { i: 'foo', p: 3 },
|
||||||
metadata: {
|
metadata: {
|
||||||
user_id: ObjectId(user_id),
|
user_id: ObjectId(userId),
|
||||||
ts: new Date(ts),
|
ts: new Date(ts),
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
comments: [
|
comments: [
|
||||||
{
|
{
|
||||||
id: ObjectId(comment_id),
|
id: ObjectId(commentId),
|
||||||
op: { c: 'foo', p: 3, t: ObjectId(thread_id) },
|
op: { c: 'foo', p: 3, t: ObjectId(threadId) },
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
it('should leave malformed ObjectIds as they are', function () {
|
it('should leave malformed ObjectIds as they are', function () {
|
||||||
const change_id = 'foo'
|
const changeId = 'foo'
|
||||||
const comment_id = 'bar'
|
const commentId = 'bar'
|
||||||
const user_id = 'baz'
|
const userId = 'baz'
|
||||||
return this.RangeManager.jsonRangesToMongo({
|
return this.RangeManager.jsonRangesToMongo({
|
||||||
changes: [
|
changes: [
|
||||||
{
|
{
|
||||||
id: change_id,
|
id: changeId,
|
||||||
metadata: {
|
metadata: {
|
||||||
user_id,
|
user_id: userId,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
comments: [
|
comments: [
|
||||||
{
|
{
|
||||||
id: comment_id,
|
id: commentId,
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
}).should.deep.equal({
|
}).should.deep.equal({
|
||||||
changes: [
|
changes: [
|
||||||
{
|
{
|
||||||
id: change_id,
|
id: changeId,
|
||||||
metadata: {
|
metadata: {
|
||||||
user_id,
|
user_id: userId,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
comments: [
|
comments: [
|
||||||
{
|
{
|
||||||
id: comment_id,
|
id: commentId,
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
|
||||||
return it('should be consistent when transformed through json -> mongo -> json', function () {
|
return it('should be consistent when transformed through json -> mongo -> json', function () {
|
||||||
const change_id = ObjectId().toString()
|
const changeId = ObjectId().toString()
|
||||||
const comment_id = ObjectId().toString()
|
const commentId = ObjectId().toString()
|
||||||
const user_id = ObjectId().toString()
|
const userId = ObjectId().toString()
|
||||||
const thread_id = ObjectId().toString()
|
const threadId = ObjectId().toString()
|
||||||
const ts = new Date().toJSON()
|
const ts = new Date().toJSON()
|
||||||
const ranges1 = {
|
const ranges1 = {
|
||||||
changes: [
|
changes: [
|
||||||
{
|
{
|
||||||
id: change_id,
|
id: changeId,
|
||||||
op: { i: 'foo', p: 3 },
|
op: { i: 'foo', p: 3 },
|
||||||
metadata: {
|
metadata: {
|
||||||
user_id,
|
user_id: userId,
|
||||||
ts,
|
ts,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
comments: [
|
comments: [
|
||||||
{
|
{
|
||||||
id: comment_id,
|
id: commentId,
|
||||||
op: { c: 'foo', p: 3, t: thread_id },
|
op: { c: 'foo', p: 3, t: threadId },
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
}
|
}
|
||||||
const ranges1_copy = JSON.parse(JSON.stringify(ranges1)) // jsonRangesToMongo modifies in place
|
const ranges1Copy = JSON.parse(JSON.stringify(ranges1)) // jsonRangesToMongo modifies in place
|
||||||
const ranges2 = JSON.parse(
|
const ranges2 = JSON.parse(
|
||||||
JSON.stringify(this.RangeManager.jsonRangesToMongo(ranges1_copy))
|
JSON.stringify(this.RangeManager.jsonRangesToMongo(ranges1Copy))
|
||||||
)
|
)
|
||||||
return ranges1.should.deep.equal(ranges2)
|
return ranges1.should.deep.equal(ranges2)
|
||||||
})
|
})
|
||||||
|
|
Loading…
Reference in a new issue