mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-21 20:47:08 -05:00
Merge pull request #4352 from overleaf/sk-add-ce-sp-migration-deleted-files
Add migrations for CE/SP, run scripts for deleted files and docs GitOrigin-RevId: e425a8d929a4aefdc6f61cff636fb5a1b4c67dce
This commit is contained in:
parent
397a3d97df
commit
99fe2eca51
9 changed files with 275 additions and 114 deletions
|
@ -0,0 +1,14 @@
|
||||||
|
const runScript = require('../scripts/back_fill_deleted_files.js')
|
||||||
|
|
||||||
|
exports.tags = ['server-ce', 'server-pro', 'saas']
|
||||||
|
|
||||||
|
exports.migrate = async client => {
|
||||||
|
const options = {
|
||||||
|
performCleanup: true,
|
||||||
|
letUserDoubleCheckInputsFor: 10,
|
||||||
|
fixPartialInserts: true,
|
||||||
|
}
|
||||||
|
await runScript(options)
|
||||||
|
}
|
||||||
|
|
||||||
|
exports.rollback = async client => {}
|
|
@ -0,0 +1,13 @@
|
||||||
|
const runScript = require('../scripts/back_fill_doc_name_for_deleted_docs.js')
|
||||||
|
|
||||||
|
exports.tags = ['server-ce', 'server-pro', 'saas']
|
||||||
|
|
||||||
|
exports.migrate = async client => {
|
||||||
|
const options = {
|
||||||
|
performCleanup: true,
|
||||||
|
letUserDoubleCheckInputsFor: 10,
|
||||||
|
}
|
||||||
|
await runScript(options)
|
||||||
|
}
|
||||||
|
|
||||||
|
exports.rollback = async client => {}
|
|
@ -0,0 +1,23 @@
|
||||||
|
const runScript = require('../scripts/back_fill_dummy_doc_meta.js')
|
||||||
|
|
||||||
|
exports.tags = ['server-ce', 'server-pro', 'saas']
|
||||||
|
|
||||||
|
exports.migrate = async client => {
|
||||||
|
const { db } = client
|
||||||
|
const [firstProject] = await db.projects
|
||||||
|
.find()
|
||||||
|
.sort({ _id: 1 })
|
||||||
|
.limit(1)
|
||||||
|
.toArray()
|
||||||
|
if (!firstProject) {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
const options = {
|
||||||
|
firstProjectId: firstProject._id,
|
||||||
|
performCleanup: true,
|
||||||
|
letUserDoubleCheckInputsFor: 10,
|
||||||
|
}
|
||||||
|
await runScript(options)
|
||||||
|
}
|
||||||
|
|
||||||
|
exports.rollback = async client => {}
|
|
@ -7,7 +7,11 @@ const {
|
||||||
|
|
||||||
class Adapter {
|
class Adapter {
|
||||||
constructor(params) {
|
constructor(params) {
|
||||||
if (!(process.argv.includes('-t') || process.argv.includes('--tags'))) {
|
if (
|
||||||
|
!process.env.SKIP_TAG_CHECK &&
|
||||||
|
!process.argv.includes('create') &&
|
||||||
|
!(process.argv.includes('-t') || process.argv.includes('--tags'))
|
||||||
|
) {
|
||||||
console.error("ERROR: must pass tags using '-t' or '--tags', exiting")
|
console.error("ERROR: must pass tags using '-t' or '--tags', exiting")
|
||||||
process.exit(1)
|
process.exit(1)
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,58 +1,62 @@
|
||||||
const WRITE_CONCURRENCY = parseInt(process.env.WRITE_CONCURRENCY, 10) || 10
|
|
||||||
|
|
||||||
const { batchedUpdate } = require('./helpers/batchedUpdate')
|
const { batchedUpdate } = require('./helpers/batchedUpdate')
|
||||||
const { promiseMapWithLimit, promisify } = require('../app/src/util/promises')
|
const { promiseMapWithLimit, promisify } = require('../app/src/util/promises')
|
||||||
const { db } = require('../app/src/infrastructure/mongodb')
|
const { db } = require('../app/src/infrastructure/mongodb')
|
||||||
const sleep = promisify(setTimeout)
|
const sleep = promisify(setTimeout)
|
||||||
|
const _ = require('lodash')
|
||||||
|
|
||||||
const PERFORM_CLEANUP = process.argv.includes('--perform-cleanup')
|
async function main(options) {
|
||||||
const FIX_PARTIAL_INSERTS = process.argv.includes('--fix-partial-inserts')
|
if (!options) {
|
||||||
const LET_USER_DOUBLE_CHECK_INPUTS_FOR = parseInt(
|
options = {}
|
||||||
process.env.LET_USER_DOUBLE_CHECK_INPUTS_FOR || 10 * 1000,
|
}
|
||||||
10
|
_.defaults(options, {
|
||||||
)
|
writeConcurrency: parseInt(process.env.WRITE_CONCURRENCY, 10) || 10,
|
||||||
|
performCleanup: process.argv.includes('--perform-cleanup'),
|
||||||
|
fixPartialInserts: process.argv.includes('--fix-partial-inserts'),
|
||||||
|
letUserDoubleCheckInputsFor: parseInt(
|
||||||
|
process.env.LET_USER_DOUBLE_CHECK_INPUTS_FOR || 10 * 1000,
|
||||||
|
10
|
||||||
|
),
|
||||||
|
})
|
||||||
|
|
||||||
async function main() {
|
await letUserDoubleCheckInputs(options)
|
||||||
await letUserDoubleCheckInputs()
|
|
||||||
|
|
||||||
await batchedUpdate(
|
await batchedUpdate(
|
||||||
'projects',
|
'projects',
|
||||||
// array is not empty ~ array has one item
|
// array is not empty ~ array has one item
|
||||||
{ 'deletedFiles.0': { $exists: true } },
|
{ 'deletedFiles.0': { $exists: true } },
|
||||||
processBatch,
|
async (x, projects) => {
|
||||||
|
await processBatch(x, projects, options)
|
||||||
|
},
|
||||||
{ _id: 1, deletedFiles: 1 }
|
{ _id: 1, deletedFiles: 1 }
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
main()
|
async function processBatch(_, projects, options) {
|
||||||
.then(() => {
|
await promiseMapWithLimit(
|
||||||
process.exit(0)
|
options.writeConcurrency,
|
||||||
})
|
projects,
|
||||||
.catch(error => {
|
async project => {
|
||||||
console.error({ error })
|
await processProject(project, options)
|
||||||
process.exit(1)
|
}
|
||||||
})
|
)
|
||||||
|
|
||||||
async function processBatch(_, projects) {
|
|
||||||
await promiseMapWithLimit(WRITE_CONCURRENCY, projects, processProject)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async function processProject(project) {
|
async function processProject(project, options) {
|
||||||
await backFillFiles(project)
|
await backFillFiles(project, options)
|
||||||
|
|
||||||
if (PERFORM_CLEANUP) {
|
if (options.performCleanup) {
|
||||||
await cleanupProject(project)
|
await cleanupProject(project)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async function backFillFiles(project) {
|
async function backFillFiles(project, options) {
|
||||||
const projectId = project._id
|
const projectId = project._id
|
||||||
filterDuplicatesInPlace(project)
|
filterDuplicatesInPlace(project)
|
||||||
project.deletedFiles.forEach(file => {
|
project.deletedFiles.forEach(file => {
|
||||||
file.projectId = projectId
|
file.projectId = projectId
|
||||||
})
|
})
|
||||||
|
|
||||||
if (FIX_PARTIAL_INSERTS) {
|
if (options.fixPartialInserts) {
|
||||||
await fixPartialInserts(project)
|
await fixPartialInserts(project)
|
||||||
} else {
|
} else {
|
||||||
await db.deletedFiles.insertMany(project.deletedFiles)
|
await db.deletedFiles.insertMany(project.deletedFiles)
|
||||||
|
@ -98,8 +102,8 @@ async function cleanupProject(project) {
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
async function letUserDoubleCheckInputs() {
|
async function letUserDoubleCheckInputs(options) {
|
||||||
if (PERFORM_CLEANUP) {
|
if (options.performCleanup) {
|
||||||
console.error('BACK FILLING AND PERFORMING CLEANUP')
|
console.error('BACK FILLING AND PERFORMING CLEANUP')
|
||||||
} else {
|
} else {
|
||||||
console.error(
|
console.error(
|
||||||
|
@ -108,8 +112,21 @@ async function letUserDoubleCheckInputs() {
|
||||||
}
|
}
|
||||||
console.error(
|
console.error(
|
||||||
'Waiting for you to double check inputs for',
|
'Waiting for you to double check inputs for',
|
||||||
LET_USER_DOUBLE_CHECK_INPUTS_FOR,
|
options.letUserDoubleCheckInputsFor,
|
||||||
'ms'
|
'ms'
|
||||||
)
|
)
|
||||||
await sleep(LET_USER_DOUBLE_CHECK_INPUTS_FOR)
|
await sleep(options.letUserDoubleCheckInputsFor)
|
||||||
|
}
|
||||||
|
|
||||||
|
module.exports = main
|
||||||
|
|
||||||
|
if (require.main === module) {
|
||||||
|
main()
|
||||||
|
.then(() => {
|
||||||
|
process.exit(0)
|
||||||
|
})
|
||||||
|
.catch(error => {
|
||||||
|
console.error({ error })
|
||||||
|
process.exit(1)
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,46 +1,50 @@
|
||||||
const WRITE_CONCURRENCY = parseInt(process.env.WRITE_CONCURRENCY, 10) || 10
|
|
||||||
|
|
||||||
const { batchedUpdate } = require('./helpers/batchedUpdate')
|
const { batchedUpdate } = require('./helpers/batchedUpdate')
|
||||||
const { promiseMapWithLimit, promisify } = require('../app/src/util/promises')
|
const { promiseMapWithLimit, promisify } = require('../app/src/util/promises')
|
||||||
const { db } = require('../app/src/infrastructure/mongodb')
|
const { db } = require('../app/src/infrastructure/mongodb')
|
||||||
const sleep = promisify(setTimeout)
|
const sleep = promisify(setTimeout)
|
||||||
|
const _ = require('lodash')
|
||||||
|
|
||||||
const PERFORM_CLEANUP = process.argv.pop() === '--perform-cleanup'
|
async function main(options) {
|
||||||
const LET_USER_DOUBLE_CHECK_INPUTS_FOR = parseInt(
|
if (!options) {
|
||||||
process.env.LET_USER_DOUBLE_CHECK_INPUTS_FOR || 10 * 1000,
|
options = {}
|
||||||
10
|
}
|
||||||
)
|
_.defaults(options, {
|
||||||
|
writeConcurrency: parseInt(process.env.WRITE_CONCURRENCY, 10) || 10,
|
||||||
|
performCleanup: process.argv.pop() === '--perform-cleanup',
|
||||||
|
letUserDoubleCheckInputsFor: parseInt(
|
||||||
|
process.env.LET_USER_DOUBLE_CHECK_INPUTS_FOR || 10 * 1000,
|
||||||
|
10
|
||||||
|
),
|
||||||
|
})
|
||||||
|
|
||||||
async function main() {
|
await letUserDoubleCheckInputs(options)
|
||||||
await letUserDoubleCheckInputs()
|
|
||||||
|
|
||||||
await batchedUpdate(
|
await batchedUpdate(
|
||||||
'projects',
|
'projects',
|
||||||
// array is not empty ~ array has one item
|
// array is not empty ~ array has one item
|
||||||
{ 'deletedDocs.0': { $exists: true } },
|
{ 'deletedDocs.0': { $exists: true } },
|
||||||
processBatch,
|
async (_collection, projects) => {
|
||||||
|
await processBatch(_collection, projects, options)
|
||||||
|
},
|
||||||
{ _id: 1, deletedDocs: 1 }
|
{ _id: 1, deletedDocs: 1 }
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
main()
|
async function processBatch(_, projects, options) {
|
||||||
.then(() => {
|
await promiseMapWithLimit(
|
||||||
process.exit(0)
|
options.writeConcurrency,
|
||||||
})
|
projects,
|
||||||
.catch(error => {
|
async project => {
|
||||||
console.error({ error })
|
await processProject(project, options)
|
||||||
process.exit(1)
|
}
|
||||||
})
|
)
|
||||||
|
|
||||||
async function processBatch(_, projects) {
|
|
||||||
await promiseMapWithLimit(WRITE_CONCURRENCY, projects, processProject)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async function processProject(project) {
|
async function processProject(project, options) {
|
||||||
for (const doc of project.deletedDocs) {
|
for (const doc of project.deletedDocs) {
|
||||||
await backFillDoc(doc)
|
await backFillDoc(doc)
|
||||||
}
|
}
|
||||||
if (PERFORM_CLEANUP) {
|
if (options.performCleanup) {
|
||||||
await cleanupProject(project)
|
await cleanupProject(project)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -57,8 +61,8 @@ async function cleanupProject(project) {
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
async function letUserDoubleCheckInputs() {
|
async function letUserDoubleCheckInputs(options) {
|
||||||
if (PERFORM_CLEANUP) {
|
if (options.performCleanup) {
|
||||||
console.error('BACK FILLING AND PERFORMING CLEANUP')
|
console.error('BACK FILLING AND PERFORMING CLEANUP')
|
||||||
} else {
|
} else {
|
||||||
console.error(
|
console.error(
|
||||||
|
@ -67,8 +71,21 @@ async function letUserDoubleCheckInputs() {
|
||||||
}
|
}
|
||||||
console.error(
|
console.error(
|
||||||
'Waiting for you to double check inputs for',
|
'Waiting for you to double check inputs for',
|
||||||
LET_USER_DOUBLE_CHECK_INPUTS_FOR,
|
options.letUserDoubleCheckInputsFor,
|
||||||
'ms'
|
'ms'
|
||||||
)
|
)
|
||||||
await sleep(LET_USER_DOUBLE_CHECK_INPUTS_FOR)
|
await sleep(options.letUserDoubleCheckInputsFor)
|
||||||
|
}
|
||||||
|
|
||||||
|
module.exports = main
|
||||||
|
|
||||||
|
if (require.main === module) {
|
||||||
|
main()
|
||||||
|
.then(() => {
|
||||||
|
process.exit(0)
|
||||||
|
})
|
||||||
|
.catch(error => {
|
||||||
|
console.error({ error })
|
||||||
|
process.exit(1)
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
|
@ -2,45 +2,54 @@ const { promisify } = require('util')
|
||||||
const { ObjectId, ReadPreference } = require('mongodb')
|
const { ObjectId, ReadPreference } = require('mongodb')
|
||||||
const { db, waitForDb } = require('../app/src/infrastructure/mongodb')
|
const { db, waitForDb } = require('../app/src/infrastructure/mongodb')
|
||||||
const sleep = promisify(setTimeout)
|
const sleep = promisify(setTimeout)
|
||||||
|
const _ = require('lodash')
|
||||||
|
|
||||||
const NOW_IN_S = Date.now() / 1000
|
const NOW_IN_S = Date.now() / 1000
|
||||||
const ONE_WEEK_IN_S = 60 * 60 * 24 * 7
|
const ONE_WEEK_IN_S = 60 * 60 * 24 * 7
|
||||||
const TEN_SECONDS = 10 * 1000
|
const TEN_SECONDS = 10 * 1000
|
||||||
|
|
||||||
const CACHE_SIZE = parseInt(process.env.CACHE_SIZE, 10) || 100
|
|
||||||
const DRY_RUN = process.env.DRY_RUN === 'true'
|
|
||||||
if (!process.env.FIRST_PROJECT_ID) {
|
|
||||||
console.error('Set FIRST_PROJECT_ID and re-run.')
|
|
||||||
process.exit(1)
|
|
||||||
}
|
|
||||||
const FIRST_PROJECT_ID = ObjectId(process.env.FIRST_PROJECT_ID)
|
|
||||||
const INCREMENT_BY_S = parseInt(process.env.INCREMENT_BY_S, 10) || ONE_WEEK_IN_S
|
|
||||||
const BATCH_SIZE = parseInt(process.env.BATCH_SIZE, 10) || 1000
|
|
||||||
const STOP_AT_S = parseInt(process.env.STOP_AT_S, 10) || NOW_IN_S
|
|
||||||
const LET_USER_DOUBLE_CHECK_INPUTS_FOR =
|
|
||||||
parseInt(process.env.LET_USER_DOUBLE_CHECK_INPUTS_FOR, 10) || TEN_SECONDS
|
|
||||||
|
|
||||||
const DUMMY_NAME = 'unknown.tex'
|
const DUMMY_NAME = 'unknown.tex'
|
||||||
const DUMMY_TIME = new Date('2021-04-12T00:00:00.000Z')
|
const DUMMY_TIME = new Date('2021-04-12T00:00:00.000Z')
|
||||||
|
|
||||||
const LRUCache = require('lru-cache')
|
const LRUCache = require('lru-cache')
|
||||||
const deletedProjectsCache = new LRUCache({
|
let deletedProjectsCache = null
|
||||||
max: CACHE_SIZE,
|
|
||||||
})
|
|
||||||
|
|
||||||
function getSecondsFromObjectId(id) {
|
function getSecondsFromObjectId(id) {
|
||||||
return id.getTimestamp().getTime() / 1000
|
return id.getTimestamp().getTime() / 1000
|
||||||
}
|
}
|
||||||
|
|
||||||
async function main() {
|
async function main(options) {
|
||||||
await letUserDoubleCheckInputs()
|
if (!options) {
|
||||||
|
options = {}
|
||||||
|
}
|
||||||
|
_.defaults(options, {
|
||||||
|
dryRun: process.env.DRY_RUN === 'true',
|
||||||
|
cacheSize: parseInt(process.env.CACHE_SIZE, 10) || 100,
|
||||||
|
firstProjectId: ObjectId(process.env.FIRST_PROJECT_ID),
|
||||||
|
incrementByS: parseInt(process.env.INCREMENT_BY_S, 10) || ONE_WEEK_IN_S,
|
||||||
|
batchSize: parseInt(process.env.BATCH_SIZE, 10) || 1000,
|
||||||
|
stopAtS: parseInt(process.env.STOP_AT_S, 10) || NOW_IN_S,
|
||||||
|
letUserDoubleCheckInputsFor:
|
||||||
|
parseInt(process.env.LET_USER_DOUBLE_CHECK_INPUTS_FOR, 10) || TEN_SECONDS,
|
||||||
|
})
|
||||||
|
|
||||||
|
if (!options.firstProjectId) {
|
||||||
|
console.error('Set FIRST_PROJECT_ID and re-run.')
|
||||||
|
process.exit(1)
|
||||||
|
}
|
||||||
|
|
||||||
|
deletedProjectsCache = new LRUCache({
|
||||||
|
max: options.cacheSize,
|
||||||
|
})
|
||||||
|
|
||||||
|
await letUserDoubleCheckInputs(options)
|
||||||
await waitForDb()
|
await waitForDb()
|
||||||
|
|
||||||
let startId = FIRST_PROJECT_ID
|
let startId = options.firstProjectId
|
||||||
|
|
||||||
let nProcessed = 0
|
let nProcessed = 0
|
||||||
while (getSecondsFromObjectId(startId) <= STOP_AT_S) {
|
while (getSecondsFromObjectId(startId) <= options.stopAtS) {
|
||||||
const end = getSecondsFromObjectId(startId) + INCREMENT_BY_S
|
const end = getSecondsFromObjectId(startId) + options.incrementByS
|
||||||
let endId = ObjectId.createFromTime(end)
|
let endId = ObjectId.createFromTime(end)
|
||||||
const query = {
|
const query = {
|
||||||
project_id: {
|
project_id: {
|
||||||
|
@ -57,16 +66,16 @@ async function main() {
|
||||||
const docs = await db.docs
|
const docs = await db.docs
|
||||||
.find(query, { readPreference: ReadPreference.SECONDARY })
|
.find(query, { readPreference: ReadPreference.SECONDARY })
|
||||||
.project({ _id: 1, project_id: 1 })
|
.project({ _id: 1, project_id: 1 })
|
||||||
.limit(BATCH_SIZE)
|
.limit(options.batchSize)
|
||||||
.toArray()
|
.toArray()
|
||||||
|
|
||||||
if (docs.length) {
|
if (docs.length) {
|
||||||
const docIds = docs.map(doc => doc._id)
|
const docIds = docs.map(doc => doc._id)
|
||||||
console.log('Back filling dummy meta data for', JSON.stringify(docIds))
|
console.log('Back filling dummy meta data for', JSON.stringify(docIds))
|
||||||
await processBatch(docs)
|
await processBatch(docs, options)
|
||||||
nProcessed += docIds.length
|
nProcessed += docIds.length
|
||||||
|
|
||||||
if (docs.length === BATCH_SIZE) {
|
if (docs.length === options.batchSize) {
|
||||||
endId = docs[docs.length - 1].project_id
|
endId = docs[docs.length - 1].project_id
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -94,7 +103,7 @@ async function getDeletedProject(projectId) {
|
||||||
return deletedProject
|
return deletedProject
|
||||||
}
|
}
|
||||||
|
|
||||||
async function processBatch(docs) {
|
async function processBatch(docs, options) {
|
||||||
for (const doc of docs) {
|
for (const doc of docs) {
|
||||||
const { _id: docId, project_id: projectId } = doc
|
const { _id: docId, project_id: projectId } = doc
|
||||||
const deletedProject = await getDeletedProject(projectId)
|
const deletedProject = await getDeletedProject(projectId)
|
||||||
|
@ -119,42 +128,31 @@ async function processBatch(docs) {
|
||||||
} else {
|
} else {
|
||||||
console.log('Orphaned deleted doc %s (no deletedProjects entry)', docId)
|
console.log('Orphaned deleted doc %s (no deletedProjects entry)', docId)
|
||||||
}
|
}
|
||||||
if (DRY_RUN) continue
|
if (options.dryRun) continue
|
||||||
await db.docs.updateOne({ _id: docId }, { $set: { name, deletedAt } })
|
await db.docs.updateOne({ _id: docId }, { $set: { name, deletedAt } })
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async function letUserDoubleCheckInputs() {
|
async function letUserDoubleCheckInputs(options) {
|
||||||
console.error(
|
console.error('Options:', JSON.stringify(options, null, 2))
|
||||||
'Options:',
|
|
||||||
JSON.stringify(
|
|
||||||
{
|
|
||||||
BATCH_SIZE,
|
|
||||||
CACHE_SIZE,
|
|
||||||
DRY_RUN,
|
|
||||||
FIRST_PROJECT_ID,
|
|
||||||
INCREMENT_BY_S,
|
|
||||||
STOP_AT_S,
|
|
||||||
LET_USER_DOUBLE_CHECK_INPUTS_FOR,
|
|
||||||
},
|
|
||||||
null,
|
|
||||||
2
|
|
||||||
)
|
|
||||||
)
|
|
||||||
console.error(
|
console.error(
|
||||||
'Waiting for you to double check inputs for',
|
'Waiting for you to double check inputs for',
|
||||||
LET_USER_DOUBLE_CHECK_INPUTS_FOR,
|
options.letUserDoubleCheckInputsFor,
|
||||||
'ms'
|
'ms'
|
||||||
)
|
)
|
||||||
await sleep(LET_USER_DOUBLE_CHECK_INPUTS_FOR)
|
await sleep(options.letUserDoubleCheckInputsFor)
|
||||||
}
|
}
|
||||||
|
|
||||||
main()
|
module.exports = main
|
||||||
.then(() => {
|
|
||||||
console.error('Done.')
|
if (require.main === module) {
|
||||||
process.exit(0)
|
main()
|
||||||
})
|
.then(() => {
|
||||||
.catch(error => {
|
console.error('Done.')
|
||||||
console.error({ error })
|
process.exit(0)
|
||||||
process.exit(1)
|
})
|
||||||
})
|
.catch(error => {
|
||||||
|
console.error({ error })
|
||||||
|
process.exit(1)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
67
services/web/scripts/mark_migration.js
Normal file
67
services/web/scripts/mark_migration.js
Normal file
|
@ -0,0 +1,67 @@
|
||||||
|
const Adapter = require('../migrations/lib/adapter')
|
||||||
|
const fs = require('fs').promises
|
||||||
|
const path = require('path')
|
||||||
|
|
||||||
|
async function main(args) {
|
||||||
|
if (
|
||||||
|
!args ||
|
||||||
|
args.length === 0 ||
|
||||||
|
args.includes('help') ||
|
||||||
|
args.includes('--help') ||
|
||||||
|
args.includes('-h')
|
||||||
|
) {
|
||||||
|
console.log('')
|
||||||
|
console.log('usage: node ./scripts/mark_migration.js migration state')
|
||||||
|
console.log('')
|
||||||
|
console.log(' migration: name of migration file')
|
||||||
|
console.log(' state: executed | unexecuted')
|
||||||
|
console.log('')
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
const migration = args[0]
|
||||||
|
if (!migration) {
|
||||||
|
throw new Error('Error: migration must be supplied')
|
||||||
|
}
|
||||||
|
const state = args[1]
|
||||||
|
if (!state) {
|
||||||
|
throw new Error('Error: migration state must be supplied')
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
await fs.access(path.join(__dirname, '../migrations', `${migration}.js`))
|
||||||
|
} catch (err) {
|
||||||
|
throw new Error(
|
||||||
|
`Error: migration ${migration} does not exist on disk: ${err}`
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
console.log(`Marking ${migration} as ${state}`)
|
||||||
|
|
||||||
|
process.env.SKIP_TAG_CHECK = 'true'
|
||||||
|
const adapter = new Adapter()
|
||||||
|
await adapter.connect()
|
||||||
|
switch (state) {
|
||||||
|
case 'executed':
|
||||||
|
await adapter.markExecuted(migration)
|
||||||
|
break
|
||||||
|
case 'unexecuted':
|
||||||
|
await adapter.unmarkExecuted(migration)
|
||||||
|
break
|
||||||
|
default:
|
||||||
|
throw new Error(`invalid state "${state}"`)
|
||||||
|
}
|
||||||
|
console.log('Done')
|
||||||
|
}
|
||||||
|
|
||||||
|
if (require.main === module) {
|
||||||
|
const args = process.argv.slice(2)
|
||||||
|
main(args)
|
||||||
|
.then(() => {
|
||||||
|
process.exit(0)
|
||||||
|
})
|
||||||
|
.catch(err => {
|
||||||
|
console.error(err)
|
||||||
|
process.exit(1)
|
||||||
|
})
|
||||||
|
}
|
|
@ -188,7 +188,15 @@ describe('BackFillDummyDocMeta', function () {
|
||||||
'',
|
'',
|
||||||
])
|
])
|
||||||
expect(stdErr.filter(filterOutput)).to.deep.equal([
|
expect(stdErr.filter(filterOutput)).to.deep.equal([
|
||||||
...`Options: ${JSON.stringify(options, null, 2)}`.split('\n'),
|
`Options: {`,
|
||||||
|
` "dryRun": ${options.DRY_RUN},`,
|
||||||
|
` "cacheSize": ${options.CACHE_SIZE},`,
|
||||||
|
` "firstProjectId": "${options.FIRST_PROJECT_ID}",`,
|
||||||
|
` "incrementByS": ${options.INCREMENT_BY_S},`,
|
||||||
|
` "batchSize": ${options.BATCH_SIZE},`,
|
||||||
|
` "stopAtS": ${options.STOP_AT_S},`,
|
||||||
|
` "letUserDoubleCheckInputsFor": ${options.LET_USER_DOUBLE_CHECK_INPUTS_FOR}`,
|
||||||
|
'}',
|
||||||
'Waiting for you to double check inputs for 1 ms',
|
'Waiting for you to double check inputs for 1 ms',
|
||||||
`Processed 1 until ${getObjectIdFromDate('2021-04-02T00:00:00.000Z')}`,
|
`Processed 1 until ${getObjectIdFromDate('2021-04-02T00:00:00.000Z')}`,
|
||||||
`Processed 2 until ${getObjectIdFromDate('2021-04-03T00:00:00.000Z')}`,
|
`Processed 2 until ${getObjectIdFromDate('2021-04-03T00:00:00.000Z')}`,
|
||||||
|
|
Loading…
Reference in a new issue