mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-21 20:47:08 -05:00
a7517eefcb
[web] populate db with collections on import, ahead of waitForDb() call GitOrigin-RevId: 7eb4cd61c2052187acd9947d7060f54d9822d314
162 lines
4.5 KiB
JavaScript
162 lines
4.5 KiB
JavaScript
import fs from 'fs'
|
|
import minimist from 'minimist'
|
|
import { ObjectId } from '../app/src/infrastructure/mongodb.js'
|
|
import DocstoreManager from '../app/src/Features/Docstore/DocstoreManager.js'
|
|
import FileStoreHandler from '../app/src/Features/FileStore/FileStoreHandler.js'
|
|
import FileWriter from '../app/src/infrastructure/FileWriter.js'
|
|
import ProjectEntityMongoUpdateHandler from '../app/src/Features/Project/ProjectEntityMongoUpdateHandler.js'
|
|
import ProjectLocator from '../app/src/Features/Project/ProjectLocator.js'
|
|
import RedisWrapper from '@overleaf/redis-wrapper'
|
|
import Settings from '@overleaf/settings'
|
|
|
|
const opts = parseArgs()
|
|
const redis = RedisWrapper.createClient(Settings.redis.web)
|
|
|
|
function parseArgs() {
|
|
const args = minimist(process.argv.slice(2), {
|
|
boolean: ['commit', 'ignore-ranges'],
|
|
})
|
|
|
|
const projectIds = args._
|
|
if (projectIds.length === 0) {
|
|
console.log(`Usage: ${process.argv[1]} [OPTS] PROJECT_ID
|
|
|
|
Options:
|
|
--commit Actually convert oversized docs to binary files
|
|
--max-doc-size Size over which docs are converted to binary files
|
|
--ignore-ranges Convert docs even if they contain ranges
|
|
`)
|
|
process.exit(0)
|
|
}
|
|
|
|
const commit = args.commit
|
|
const ignoreRanges = args['ignore-ranges']
|
|
const maxDocSize = args['max-doc-size']
|
|
? parseInt(args['max-doc-size'], 10)
|
|
: 2 * 1024 * 1024
|
|
|
|
return { projectIds, commit, ignoreRanges, maxDocSize }
|
|
}
|
|
|
|
async function main() {
|
|
for (const projectId of opts.projectIds) {
|
|
await processProject(projectId)
|
|
}
|
|
if (!opts.commit) {
|
|
console.log('This was a dry run. Re-run with --commit to apply changes')
|
|
}
|
|
}
|
|
|
|
async function processProject(projectId) {
|
|
const docIds = await getDocIds(projectId)
|
|
for (const docId of docIds) {
|
|
await processDoc(projectId, docId)
|
|
}
|
|
}
|
|
|
|
async function processDoc(projectId, docId) {
|
|
const doc = await getDoc(projectId, docId)
|
|
const size = doc.lines.reduce((sum, line) => sum + line.length + 1, 0)
|
|
if (size > opts.maxDocSize) {
|
|
if (
|
|
!opts.ignoreRanges &&
|
|
((doc.ranges.comments && doc.ranges.comments.length > 0) ||
|
|
(doc.ranges.changes && doc.ranges.changes.length > 0))
|
|
) {
|
|
console.log(
|
|
`Skipping doc ${doc.path} in project ${projectId} because it has ranges`
|
|
)
|
|
return
|
|
}
|
|
console.log(
|
|
`Converting doc ${doc.path} in project ${projectId} to binary (${size} bytes)`
|
|
)
|
|
if (opts.commit) {
|
|
const fileRef = await sendDocToFilestore(projectId, doc)
|
|
await ProjectEntityMongoUpdateHandler.promises.replaceDocWithFile(
|
|
new ObjectId(projectId),
|
|
new ObjectId(docId),
|
|
fileRef
|
|
)
|
|
await deleteDocFromMongo(projectId, doc)
|
|
await deleteDocFromRedis(projectId, docId)
|
|
}
|
|
}
|
|
}
|
|
|
|
async function getDocIds(projectId) {
|
|
const docIds = await redis.smembers(`DocsIn:{${projectId}}`)
|
|
return docIds
|
|
}
|
|
|
|
async function getDoc(projectId, docId) {
|
|
const lines = await redis.get(`doclines:{${docId}}`)
|
|
const ranges = await redis.get(`Ranges:{${docId}}`)
|
|
const { path } = await ProjectLocator.promises.findElement({
|
|
project_id: projectId,
|
|
element_id: docId,
|
|
type: 'doc',
|
|
})
|
|
return {
|
|
id: docId,
|
|
lines: JSON.parse(lines),
|
|
ranges: ranges ? JSON.parse(ranges) : {},
|
|
path: path.fileSystem,
|
|
}
|
|
}
|
|
|
|
async function sendDocToFilestore(projectId, doc) {
|
|
const basename = doc.path.split('/').pop()
|
|
const tmpFilePath = await FileWriter.promises.writeLinesToDisk(
|
|
projectId,
|
|
doc.lines
|
|
)
|
|
try {
|
|
const { fileRef } = await FileStoreHandler.promises.uploadFileFromDisk(
|
|
projectId,
|
|
{ name: basename, rev: doc.version + 1 },
|
|
tmpFilePath
|
|
)
|
|
return fileRef
|
|
} finally {
|
|
fs.promises.unlink(tmpFilePath)
|
|
}
|
|
}
|
|
|
|
async function deleteDocFromMongo(projectId, doc) {
|
|
const basename = doc.path.split('/').pop()
|
|
const deletedAt = new Date()
|
|
await DocstoreManager.promises.deleteDoc(
|
|
projectId,
|
|
doc.id,
|
|
basename,
|
|
deletedAt
|
|
)
|
|
}
|
|
|
|
async function deleteDocFromRedis(projectId, docId) {
|
|
await redis.del(
|
|
`Blocking:{${docId}}`,
|
|
`doclines:{${docId}}`,
|
|
`DocOps:{${docId}}`,
|
|
`DocVersion:{${docId}}`,
|
|
`DocHash:{${docId}}`,
|
|
`ProjectId:{${docId}}`,
|
|
`Ranges:{${docId}}`,
|
|
`UnflushedTime:{${docId}}`,
|
|
`Pathname:{${docId}}`,
|
|
`ProjectHistoryId:{${docId}}`,
|
|
`PendingUpdates:{${docId}}`,
|
|
`lastUpdatedAt:{${docId}}`,
|
|
`lastUpdatedBy:{${docId}}`
|
|
)
|
|
await redis.srem(`DocsIn:{${projectId}}`, projectId)
|
|
}
|
|
|
|
try {
|
|
await main()
|
|
process.exit(0)
|
|
} catch (error) {
|
|
console.error(error)
|
|
process.exit(1)
|
|
}
|