mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-07 20:31:06 -05:00
e73fdfba63
History ranges migration script - second attempt GitOrigin-RevId: 60a2c04e2a72e76a58e9e179fefc4186a96fde32
1097 lines
32 KiB
JavaScript
1097 lines
32 KiB
JavaScript
// @ts-check
|
|
|
|
import _ from 'lodash'
|
|
import { callbackify, promisify } from 'util'
|
|
import { callbackifyMultiResult } from '@overleaf/promise-utils'
|
|
import Settings from '@overleaf/settings'
|
|
import logger from '@overleaf/logger'
|
|
import Metrics from '@overleaf/metrics'
|
|
import OError from '@overleaf/o-error'
|
|
import { File, Range } from 'overleaf-editor-core'
|
|
import { SyncError } from './Errors.js'
|
|
import { db, ObjectId } from './mongodb.js'
|
|
import * as SnapshotManager from './SnapshotManager.js'
|
|
import * as LockManager from './LockManager.js'
|
|
import * as UpdateTranslator from './UpdateTranslator.js'
|
|
import * as UpdateCompressor from './UpdateCompressor.js'
|
|
import * as WebApiManager from './WebApiManager.js'
|
|
import * as ErrorRecorder from './ErrorRecorder.js'
|
|
import * as RedisManager from './RedisManager.js'
|
|
import * as HistoryStoreManager from './HistoryStoreManager.js'
|
|
import * as HashManager from './HashManager.js'
|
|
import { isInsert, isDelete } from './Utils.js'
|
|
|
|
/**
|
|
* @typedef {import('overleaf-editor-core').Comment} HistoryComment
|
|
* @typedef {import('overleaf-editor-core').TrackedChange} HistoryTrackedChange
|
|
* @typedef {import('./types').Comment} Comment
|
|
* @typedef {import('./types').Entity} Entity
|
|
* @typedef {import('./types').ResyncDocContentUpdate} ResyncDocContentUpdate
|
|
* @typedef {import('./types').RetainOp} RetainOp
|
|
* @typedef {import('./types').TrackedChange} TrackedChange
|
|
* @typedef {import('./types').TrackedChangeTransition} TrackedChangeTransition
|
|
* @typedef {import('./types').TrackingDirective} TrackingDirective
|
|
* @typedef {import('./types').TrackingType} TrackingType
|
|
* @typedef {import('./types').Update} Update
|
|
*/
|
|
const MAX_RESYNC_HISTORY_RECORDS = 100 // keep this many records of previous resyncs
|
|
const EXPIRE_RESYNC_HISTORY_INTERVAL_MS = 90 * 24 * 3600 * 1000 // 90 days
|
|
|
|
const keys = Settings.redis.lock.key_schema
|
|
|
|
// db.projectHistorySyncState.ensureIndex({expiresAt: 1}, {expireAfterSeconds: 0, background: true})
|
|
// To add expiresAt field to existing entries in collection (choose a suitable future expiry date):
|
|
// db.projectHistorySyncState.updateMany({resyncProjectStructure: false, resyncDocContents: [], expiresAt: {$exists:false}}, {$set: {expiresAt: new Date("2019-07-01")}})
|
|
|
|
async function startResync(projectId, options = {}) {
|
|
// We have three options here
|
|
//
|
|
// 1. If we update mongo before making the call to web then there's a
|
|
// chance we ignore all updates indefinitely (there's no foolproff way
|
|
// to undo the change in mongo)
|
|
//
|
|
// 2. If we make the call to web first then there is a small race condition
|
|
// where we could process the sync update and then only update mongo
|
|
// after, causing all updates to be ignored from then on
|
|
//
|
|
// 3. We can wrap everything in a project lock
|
|
Metrics.inc('project_history_resync')
|
|
try {
|
|
await LockManager.promises.runWithLock(
|
|
keys.projectHistoryLock({ project_id: projectId }),
|
|
async extendLock => {
|
|
await _startResyncWithoutLock(projectId, options)
|
|
}
|
|
)
|
|
} catch (error) {
|
|
// record error in starting sync ("sync ongoing")
|
|
try {
|
|
await ErrorRecorder.promises.record(projectId, -1, error)
|
|
} catch (err) {
|
|
// swallow any error thrown by ErrorRecorder.record()
|
|
}
|
|
throw error
|
|
}
|
|
}
|
|
|
|
async function startHardResync(projectId, options = {}) {
|
|
Metrics.inc('project_history_hard_resync')
|
|
try {
|
|
await LockManager.promises.runWithLock(
|
|
keys.projectHistoryLock({ project_id: projectId }),
|
|
async extendLock => {
|
|
await clearResyncState(projectId)
|
|
await RedisManager.promises.clearFirstOpTimestamp(projectId)
|
|
await RedisManager.promises.destroyDocUpdatesQueue(projectId)
|
|
await _startResyncWithoutLock(projectId, options)
|
|
}
|
|
)
|
|
} catch (error) {
|
|
// record error in starting sync ("sync ongoing")
|
|
await ErrorRecorder.promises.record(projectId, -1, error)
|
|
throw error
|
|
}
|
|
}
|
|
|
|
async function _startResyncWithoutLock(projectId, options) {
|
|
await ErrorRecorder.promises.recordSyncStart(projectId)
|
|
|
|
const syncState = await _getResyncState(projectId)
|
|
if (syncState.isSyncOngoing()) {
|
|
throw new OError('sync ongoing')
|
|
}
|
|
syncState.setOrigin(options.origin || { kind: 'history-resync' })
|
|
syncState.startProjectStructureSync()
|
|
|
|
const webOpts = {}
|
|
if (options.historyRangesMigration) {
|
|
webOpts.historyRangesMigration = options.historyRangesMigration
|
|
}
|
|
await WebApiManager.promises.requestResync(projectId, webOpts)
|
|
await setResyncState(projectId, syncState)
|
|
}
|
|
|
|
async function _getResyncState(projectId) {
|
|
const rawSyncState = await db.projectHistorySyncState.findOne({
|
|
project_id: new ObjectId(projectId.toString()),
|
|
})
|
|
const syncState = SyncState.fromRaw(projectId, rawSyncState)
|
|
return syncState
|
|
}
|
|
|
|
async function setResyncState(projectId, syncState) {
|
|
// skip if syncState is null (i.e. unchanged)
|
|
if (syncState == null) {
|
|
return
|
|
}
|
|
const update = {
|
|
$set: syncState.toRaw(),
|
|
$push: {
|
|
history: {
|
|
$each: [{ syncState: syncState.toRaw(), timestamp: new Date() }],
|
|
$position: 0,
|
|
$slice: MAX_RESYNC_HISTORY_RECORDS,
|
|
},
|
|
},
|
|
$currentDate: { lastUpdated: true },
|
|
}
|
|
|
|
// handle different cases
|
|
if (syncState.isSyncOngoing()) {
|
|
// starting a new sync; prevent the entry expiring while sync is in ongoing
|
|
update.$inc = { resyncCount: 1 }
|
|
update.$unset = { expiresAt: true }
|
|
} else {
|
|
// successful completion of existing sync; set the entry to expire in the
|
|
// future
|
|
update.$set.expiresAt = new Date(
|
|
Date.now() + EXPIRE_RESYNC_HISTORY_INTERVAL_MS
|
|
)
|
|
}
|
|
|
|
// apply the update
|
|
await db.projectHistorySyncState.updateOne(
|
|
{ project_id: new ObjectId(projectId) },
|
|
update,
|
|
{ upsert: true }
|
|
)
|
|
}
|
|
|
|
async function clearResyncState(projectId) {
|
|
await db.projectHistorySyncState.deleteOne({
|
|
project_id: new ObjectId(projectId.toString()),
|
|
})
|
|
}
|
|
|
|
async function skipUpdatesDuringSync(projectId, updates) {
|
|
const syncState = await _getResyncState(projectId)
|
|
if (!syncState.isSyncOngoing()) {
|
|
logger.debug({ projectId }, 'not skipping updates: no resync in progress')
|
|
// don't return syncState when unchanged
|
|
return { updates, syncState: null }
|
|
}
|
|
|
|
const filteredUpdates = []
|
|
|
|
for (const update of updates) {
|
|
syncState.updateState(update)
|
|
const shouldSkipUpdate = syncState.shouldSkipUpdate(update)
|
|
if (!shouldSkipUpdate) {
|
|
filteredUpdates.push(update)
|
|
} else {
|
|
logger.debug({ projectId, update }, 'skipping update due to resync')
|
|
}
|
|
}
|
|
return { updates: filteredUpdates, syncState }
|
|
}
|
|
|
|
async function expandSyncUpdates(
|
|
projectId,
|
|
projectHistoryId,
|
|
updates,
|
|
extendLock
|
|
) {
|
|
const areSyncUpdatesQueued =
|
|
_.some(updates, 'resyncProjectStructure') ||
|
|
_.some(updates, 'resyncDocContent')
|
|
if (!areSyncUpdatesQueued) {
|
|
logger.debug({ projectId }, 'no resync updates to expand')
|
|
return updates
|
|
}
|
|
|
|
const syncState = await _getResyncState(projectId)
|
|
|
|
// compute the current snapshot from the most recent chunk
|
|
const snapshotFiles = await SnapshotManager.promises.getLatestSnapshot(
|
|
projectId,
|
|
projectHistoryId
|
|
)
|
|
|
|
// check if snapshot files are valid
|
|
const invalidFiles = _.pickBy(
|
|
snapshotFiles,
|
|
(v, k) => v == null || typeof v.isEditable !== 'function'
|
|
)
|
|
if (_.size(invalidFiles) > 0) {
|
|
throw new SyncError('file is missing isEditable method', {
|
|
projectId,
|
|
invalidFiles,
|
|
})
|
|
}
|
|
|
|
const expander = new SyncUpdateExpander(
|
|
projectId,
|
|
snapshotFiles,
|
|
syncState.origin
|
|
)
|
|
|
|
// expand updates asynchronously to avoid blocking
|
|
for (const update of updates) {
|
|
await expander.expandUpdate(update)
|
|
await extendLock()
|
|
}
|
|
|
|
return expander.getExpandedUpdates()
|
|
}
|
|
|
|
class SyncState {
|
|
constructor(projectId, resyncProjectStructure, resyncDocContents, origin) {
|
|
this.projectId = projectId
|
|
this.resyncProjectStructure = resyncProjectStructure
|
|
this.resyncDocContents = resyncDocContents
|
|
this.origin = origin
|
|
}
|
|
|
|
static fromRaw(projectId, rawSyncState) {
|
|
rawSyncState = rawSyncState || {}
|
|
const resyncProjectStructure = rawSyncState.resyncProjectStructure || false
|
|
const resyncDocContents = new Set(rawSyncState.resyncDocContents || [])
|
|
const origin = rawSyncState.origin
|
|
return new SyncState(
|
|
projectId,
|
|
resyncProjectStructure,
|
|
resyncDocContents,
|
|
origin
|
|
)
|
|
}
|
|
|
|
toRaw() {
|
|
return {
|
|
resyncProjectStructure: this.resyncProjectStructure,
|
|
resyncDocContents: Array.from(this.resyncDocContents),
|
|
origin: this.origin,
|
|
}
|
|
}
|
|
|
|
updateState(update) {
|
|
if (update.resyncProjectStructure != null) {
|
|
if (!this.isProjectStructureSyncing()) {
|
|
throw new SyncError('unexpected resyncProjectStructure update', {
|
|
projectId: this.projectId,
|
|
resyncProjectStructure: this.resyncProjectStructure,
|
|
})
|
|
}
|
|
if (this.isAnyDocContentSyncing()) {
|
|
throw new SyncError('unexpected resyncDocContents update', {
|
|
projectId: this.projectId,
|
|
resyncDocContents: this.resyncDocContents,
|
|
})
|
|
}
|
|
|
|
for (const doc of update.resyncProjectStructure.docs) {
|
|
this.startDocContentSync(doc.path)
|
|
}
|
|
|
|
this.stopProjectStructureSync()
|
|
} else if (update.resyncDocContent != null) {
|
|
if (this.isProjectStructureSyncing()) {
|
|
throw new SyncError('unexpected resyncDocContent update', {
|
|
projectId: this.projectId,
|
|
resyncProjectStructure: this.resyncProjectStructure,
|
|
})
|
|
}
|
|
|
|
if (!this.isDocContentSyncing(update.path)) {
|
|
throw new SyncError('unexpected resyncDocContent update', {
|
|
projectId: this.projectId,
|
|
resyncDocContents: this.resyncDocContents,
|
|
path: update.path,
|
|
})
|
|
}
|
|
|
|
this.stopDocContentSync(update.path)
|
|
}
|
|
}
|
|
|
|
setOrigin(origin) {
|
|
this.origin = origin
|
|
}
|
|
|
|
shouldSkipUpdate(update) {
|
|
// don't skip sync updates
|
|
if (
|
|
update.resyncProjectStructure != null ||
|
|
update.resyncDocContent != null
|
|
) {
|
|
return false
|
|
}
|
|
|
|
// if syncing project structure skip update
|
|
if (this.isProjectStructureSyncing()) {
|
|
return true
|
|
}
|
|
|
|
// skip text updates for a docs being synced
|
|
if (UpdateTranslator.isTextUpdate(update)) {
|
|
if (this.isDocContentSyncing(update.meta.pathname)) {
|
|
return true
|
|
}
|
|
}
|
|
|
|
// preserve all other updates
|
|
return false
|
|
}
|
|
|
|
startProjectStructureSync() {
|
|
this.resyncProjectStructure = true
|
|
this.resyncDocContents = new Set([])
|
|
}
|
|
|
|
stopProjectStructureSync() {
|
|
this.resyncProjectStructure = false
|
|
}
|
|
|
|
stopDocContentSync(pathname) {
|
|
this.resyncDocContents.delete(pathname)
|
|
}
|
|
|
|
startDocContentSync(pathname) {
|
|
this.resyncDocContents.add(pathname)
|
|
}
|
|
|
|
isProjectStructureSyncing() {
|
|
return this.resyncProjectStructure
|
|
}
|
|
|
|
isDocContentSyncing(pathname) {
|
|
return this.resyncDocContents.has(pathname)
|
|
}
|
|
|
|
isAnyDocContentSyncing() {
|
|
return this.resyncDocContents.size > 0
|
|
}
|
|
|
|
isSyncOngoing() {
|
|
return this.isProjectStructureSyncing() || this.isAnyDocContentSyncing()
|
|
}
|
|
}
|
|
|
|
class SyncUpdateExpander {
|
|
/**
|
|
* Build a SyncUpdateExpander
|
|
*
|
|
* @param {string} projectId
|
|
* @param {Record<string, File>} snapshotFiles
|
|
* @param {string} origin
|
|
*/
|
|
constructor(projectId, snapshotFiles, origin) {
|
|
this.projectId = projectId
|
|
this.files = snapshotFiles
|
|
this.expandedUpdates = []
|
|
this.origin = origin
|
|
}
|
|
|
|
// If there's an expected *file* with the same path and either the same hash
|
|
// or no hash, treat this as not editable even if history thinks it is.
|
|
isEditable(filePath, file, expectedFiles) {
|
|
if (!file.isEditable()) {
|
|
return false
|
|
}
|
|
const fileHash = _.get(file, ['data', 'hash'])
|
|
const matchedExpectedFile = expectedFiles.some(item => {
|
|
const expectedFileHash = item._hash
|
|
if (expectedFileHash && fileHash !== expectedFileHash) {
|
|
// expected file has a hash and it doesn't match
|
|
return false
|
|
}
|
|
return UpdateTranslator._convertPathname(item.path) === filePath
|
|
})
|
|
|
|
// consider editable file in history as binary, since it matches a binary file in the project
|
|
return !matchedExpectedFile
|
|
}
|
|
|
|
/**
|
|
* @param {Update} update
|
|
*/
|
|
async expandUpdate(update) {
|
|
if ('resyncProjectStructure' in update) {
|
|
logger.debug(
|
|
{ projectId: this.projectId, update },
|
|
'expanding resyncProjectStructure update'
|
|
)
|
|
const persistedNonBinaryFileEntries = _.pickBy(this.files, (v, k) =>
|
|
this.isEditable(k, v, update.resyncProjectStructure.files)
|
|
)
|
|
const persistedNonBinaryFiles = _.map(
|
|
Object.keys(persistedNonBinaryFileEntries),
|
|
path => ({
|
|
path,
|
|
})
|
|
)
|
|
|
|
const persistedBinaryFileEntries = _.omitBy(this.files, (v, k) =>
|
|
this.isEditable(k, v, update.resyncProjectStructure.files)
|
|
)
|
|
// preserve file properties on binary files, for future comparison.
|
|
const persistedBinaryFiles = _.map(
|
|
persistedBinaryFileEntries,
|
|
(entity, key) => Object.assign({}, entity, { path: key })
|
|
)
|
|
const expectedNonBinaryFiles = _.map(
|
|
update.resyncProjectStructure.docs,
|
|
entity =>
|
|
Object.assign({}, entity, {
|
|
path: UpdateTranslator._convertPathname(entity.path),
|
|
})
|
|
)
|
|
const expectedBinaryFiles = _.map(
|
|
update.resyncProjectStructure.files,
|
|
entity =>
|
|
Object.assign({}, entity, {
|
|
path: UpdateTranslator._convertPathname(entity.path),
|
|
})
|
|
)
|
|
|
|
// We need to detect and fix consistency issues where web and
|
|
// history-store disagree on whether an entity is binary or not. Thus we
|
|
// need to remove and add the two separately.
|
|
this.queueRemoveOpsForUnexpectedFiles(
|
|
update,
|
|
expectedBinaryFiles,
|
|
persistedBinaryFiles
|
|
)
|
|
this.queueRemoveOpsForUnexpectedFiles(
|
|
update,
|
|
expectedNonBinaryFiles,
|
|
persistedNonBinaryFiles
|
|
)
|
|
this.queueAddOpsForMissingFiles(
|
|
update,
|
|
expectedBinaryFiles,
|
|
persistedBinaryFiles
|
|
)
|
|
this.queueAddOpsForMissingFiles(
|
|
update,
|
|
expectedNonBinaryFiles,
|
|
persistedNonBinaryFiles
|
|
)
|
|
this.queueUpdateForOutOfSyncBinaryFiles(
|
|
update,
|
|
expectedBinaryFiles,
|
|
persistedBinaryFiles
|
|
)
|
|
} else if ('resyncDocContent' in update) {
|
|
logger.debug(
|
|
{ projectId: this.projectId, update },
|
|
'expanding resyncDocContent update'
|
|
)
|
|
await this.expandResyncDocContentUpdate(update)
|
|
} else {
|
|
this.expandedUpdates.push(update)
|
|
}
|
|
}
|
|
|
|
getExpandedUpdates() {
|
|
return this.expandedUpdates
|
|
}
|
|
|
|
/**
|
|
* @param {Entity[]} expectedFiles
|
|
* @param {{ path: string }[]} persistedFiles
|
|
*/
|
|
queueRemoveOpsForUnexpectedFiles(update, expectedFiles, persistedFiles) {
|
|
const unexpectedFiles = _.differenceBy(
|
|
persistedFiles,
|
|
expectedFiles,
|
|
'path'
|
|
)
|
|
for (const entity of unexpectedFiles) {
|
|
update = {
|
|
pathname: entity.path,
|
|
new_pathname: '',
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
},
|
|
}
|
|
this.expandedUpdates.push(update)
|
|
Metrics.inc('project_history_resync_operation', 1, {
|
|
status: 'remove unexpected file',
|
|
})
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @param {Entity[]} expectedFiles
|
|
* @param {{ path: string }[]} persistedFiles
|
|
*/
|
|
queueAddOpsForMissingFiles(update, expectedFiles, persistedFiles) {
|
|
const missingFiles = _.differenceBy(expectedFiles, persistedFiles, 'path')
|
|
for (const entity of missingFiles) {
|
|
update = {
|
|
pathname: entity.path,
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
},
|
|
}
|
|
|
|
if ('doc' in entity) {
|
|
update.doc = entity.doc
|
|
update.docLines = ''
|
|
// we have to create a dummy entry here because later we will need the content in the diff computation
|
|
this.files[update.pathname] = File.fromString('')
|
|
} else {
|
|
update.file = entity.file
|
|
update.url = entity.url
|
|
}
|
|
|
|
this.expandedUpdates.push(update)
|
|
Metrics.inc('project_history_resync_operation', 1, {
|
|
status: 'add missing file',
|
|
})
|
|
}
|
|
}
|
|
|
|
queueUpdateForOutOfSyncBinaryFiles(update, expectedFiles, persistedFiles) {
|
|
// create a map to lookup persisted files by their path
|
|
const persistedFileMap = new Map(persistedFiles.map(x => [x.path, x]))
|
|
// now search for files with same path but different hash values
|
|
const differentFiles = expectedFiles.filter(expected => {
|
|
// check for a persisted file at the same path
|
|
const expectedPath = expected.path
|
|
const persistedFileAtSamePath = persistedFileMap.get(expectedPath)
|
|
if (!persistedFileAtSamePath) return false
|
|
// check if the persisted file at the same path has a different hash
|
|
const expectedHash = _.get(expected, '_hash')
|
|
const persistedHash = _.get(persistedFileAtSamePath, ['data', 'hash'])
|
|
const hashesPresent = expectedHash && persistedHash
|
|
return hashesPresent && persistedHash !== expectedHash
|
|
})
|
|
for (const entity of differentFiles) {
|
|
// remove the outdated persisted file
|
|
const removeUpdate = {
|
|
pathname: entity.path,
|
|
new_pathname: '',
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
},
|
|
}
|
|
this.expandedUpdates.push(removeUpdate)
|
|
// add the new file content
|
|
const addUpdate = {
|
|
pathname: entity.path,
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
},
|
|
file: entity.file,
|
|
url: entity.url,
|
|
}
|
|
this.expandedUpdates.push(addUpdate)
|
|
Metrics.inc('project_history_resync_operation', 1, {
|
|
status: 'update binary file contents',
|
|
})
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Expand a resyncDocContentUpdate
|
|
*
|
|
* @param {ResyncDocContentUpdate} update
|
|
*/
|
|
async expandResyncDocContentUpdate(update) {
|
|
const pathname = UpdateTranslator._convertPathname(update.path)
|
|
const snapshotFile = this.files[pathname]
|
|
const expectedFile = update.resyncDocContent
|
|
const expectedContent = expectedFile.content
|
|
|
|
if (!snapshotFile) {
|
|
throw new OError('unrecognised file: not in snapshot')
|
|
}
|
|
|
|
// Compare hashes to see if the persisted file matches the expected content.
|
|
// The hash of the persisted files is stored in the snapshot.
|
|
// Note getHash() returns the hash only when the persisted file has
|
|
// no changes in the snapshot, the hash is null if there are changes
|
|
// that apply to it.
|
|
let hashesMatch = false
|
|
const persistedHash = snapshotFile.getHash()
|
|
if (persistedHash != null) {
|
|
const expectedHash = HashManager._getBlobHashFromString(expectedContent)
|
|
if (persistedHash === expectedHash) {
|
|
logger.debug(
|
|
{ projectId: this.projectId, persistedHash, expectedHash },
|
|
'skipping diff because hashes match and persisted file has no ops'
|
|
)
|
|
hashesMatch = true
|
|
}
|
|
} else {
|
|
logger.debug('cannot compare hashes, will retrieve content')
|
|
}
|
|
|
|
// compute the difference between the expected and persisted content
|
|
const historyId = await WebApiManager.promises.getHistoryId(this.projectId)
|
|
const file = await snapshotFile.load(
|
|
'eager',
|
|
HistoryStoreManager.getBlobStore(historyId)
|
|
)
|
|
const persistedContent = file.getContent()
|
|
if (persistedContent == null) {
|
|
// This should not happen given that we loaded the file eagerly. We could
|
|
// probably refine the types in overleaf-editor-core so that this check
|
|
// wouldn't be necessary.
|
|
throw new Error('File was not properly loaded')
|
|
}
|
|
|
|
if (!hashesMatch) {
|
|
const expandedUpdate = await this.queueUpdateForOutOfSyncContent(
|
|
update,
|
|
pathname,
|
|
persistedContent,
|
|
expectedContent
|
|
)
|
|
if (expandedUpdate != null) {
|
|
// Adjust the ranges for the changes that have been made to the content
|
|
for (const op of expandedUpdate.op) {
|
|
if (isInsert(op)) {
|
|
file.getComments().applyInsert(new Range(op.p, op.i.length))
|
|
file.getTrackedChanges().applyInsert(op.p, op.i)
|
|
} else if (isDelete(op)) {
|
|
file.getComments().applyDelete(new Range(op.p, op.d.length))
|
|
file.getTrackedChanges().applyDelete(op.p, op.d.length)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
const persistedComments = file.getComments().toArray()
|
|
await this.queueUpdatesForOutOfSyncComments(
|
|
update,
|
|
pathname,
|
|
persistedComments
|
|
)
|
|
|
|
const persistedChanges = file.getTrackedChanges().asSorted()
|
|
await this.queueUpdatesForOutOfSyncTrackedChanges(
|
|
update,
|
|
pathname,
|
|
persistedChanges
|
|
)
|
|
}
|
|
|
|
/**
|
|
* Queue update for out of sync content
|
|
*
|
|
* @param {ResyncDocContentUpdate} update
|
|
* @param {string} pathname
|
|
* @param {string} persistedContent
|
|
* @param {string} expectedContent
|
|
*/
|
|
async queueUpdateForOutOfSyncContent(
|
|
update,
|
|
pathname,
|
|
persistedContent,
|
|
expectedContent
|
|
) {
|
|
logger.debug(
|
|
{ projectId: this.projectId, persistedContent, expectedContent },
|
|
'diffing doc contents'
|
|
)
|
|
const op = UpdateCompressor.diffAsShareJsOps(
|
|
persistedContent,
|
|
expectedContent
|
|
)
|
|
if (op.length === 0) {
|
|
return null
|
|
}
|
|
const expandedUpdate = {
|
|
doc: update.doc,
|
|
op,
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
pathname,
|
|
doc_length: persistedContent.length,
|
|
},
|
|
}
|
|
logger.debug(
|
|
{ projectId: this.projectId, diffCount: op.length },
|
|
'doc contents differ'
|
|
)
|
|
this.expandedUpdates.push(expandedUpdate)
|
|
Metrics.inc('project_history_resync_operation', 1, {
|
|
status: 'update text file contents',
|
|
})
|
|
return expandedUpdate
|
|
}
|
|
|
|
/**
|
|
* Queue updates for out of sync comments
|
|
*
|
|
* @param {ResyncDocContentUpdate} update
|
|
* @param {string} pathname
|
|
* @param {HistoryComment[]} persistedComments
|
|
*/
|
|
async queueUpdatesForOutOfSyncComments(update, pathname, persistedComments) {
|
|
const expectedContent = update.resyncDocContent.content
|
|
const expectedComments = update.resyncDocContent.ranges?.comments ?? []
|
|
const resolvedCommentIds = new Set(
|
|
update.resyncDocContent.resolvedCommentIds ?? []
|
|
)
|
|
const expectedCommentsById = new Map(
|
|
expectedComments.map(comment => [comment.id, comment])
|
|
)
|
|
const persistedCommentsById = new Map(
|
|
persistedComments.map(comment => [comment.id, comment])
|
|
)
|
|
|
|
// Delete any persisted comment that is not in the expected comment list.
|
|
for (const persistedComment of persistedComments) {
|
|
if (!expectedCommentsById.has(persistedComment.id)) {
|
|
this.expandedUpdates.push({
|
|
pathname,
|
|
deleteComment: persistedComment.id,
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
},
|
|
})
|
|
}
|
|
}
|
|
|
|
for (const expectedComment of expectedComments) {
|
|
const persistedComment = persistedCommentsById.get(expectedComment.id)
|
|
const expectedCommentResolved = resolvedCommentIds.has(expectedComment.id)
|
|
if (
|
|
persistedComment != null &&
|
|
commentRangesAreInSync(persistedComment, expectedComment)
|
|
) {
|
|
if (expectedCommentResolved === persistedComment.resolved) {
|
|
// Both comments are identical; do nothing
|
|
} else {
|
|
// Only the resolved state differs
|
|
this.expandedUpdates.push({
|
|
pathname,
|
|
commentId: expectedComment.id,
|
|
resolved: expectedCommentResolved,
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
},
|
|
})
|
|
}
|
|
} else {
|
|
const op = { ...expectedComment.op, resolved: expectedCommentResolved }
|
|
// New comment or ranges differ
|
|
this.expandedUpdates.push({
|
|
doc: update.doc,
|
|
op: [op],
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
pathname,
|
|
doc_length: expectedContent.length,
|
|
},
|
|
})
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Queue updates for out of sync tracked changes
|
|
*
|
|
* @param {ResyncDocContentUpdate} update
|
|
* @param {string} pathname
|
|
* @param {readonly HistoryTrackedChange[]} persistedChanges
|
|
*/
|
|
async queueUpdatesForOutOfSyncTrackedChanges(
|
|
update,
|
|
pathname,
|
|
persistedChanges
|
|
) {
|
|
const expectedChanges = update.resyncDocContent.ranges?.changes ?? []
|
|
const expectedContent = update.resyncDocContent.content
|
|
|
|
/**
|
|
* A cursor on the expected content
|
|
*/
|
|
let cursor = 0
|
|
|
|
/**
|
|
* The persisted tracking at cursor
|
|
*
|
|
* @type {TrackingDirective}
|
|
*/
|
|
let persistedTracking = { type: 'none' }
|
|
|
|
/**
|
|
* The expected tracking at cursor
|
|
*
|
|
* @type {TrackingDirective}
|
|
*/
|
|
let expectedTracking = { type: 'none' }
|
|
|
|
/**
|
|
* The retain ops for the update
|
|
*
|
|
* @type {RetainOp[]}
|
|
*/
|
|
const ops = []
|
|
|
|
/**
|
|
* The retain op being built
|
|
*
|
|
* @type {RetainOp | null}
|
|
*/
|
|
let currentOp = null
|
|
|
|
for (const transition of getTrackedChangesTransitions(
|
|
persistedChanges,
|
|
expectedChanges,
|
|
expectedContent.length
|
|
)) {
|
|
if (transition.pos > cursor) {
|
|
// The next transition will move the cursor. Decide what to do with the interval.
|
|
if (trackingDirectivesEqual(expectedTracking, persistedTracking)) {
|
|
// Expected tracking and persisted tracking are in sync. Emit the
|
|
// current op and skip this interval.
|
|
if (currentOp != null) {
|
|
ops.push(currentOp)
|
|
currentOp = null
|
|
}
|
|
} else {
|
|
// Expected tracking and persisted tracking are different.
|
|
const retainedText = expectedContent.slice(cursor, transition.pos)
|
|
if (
|
|
currentOp?.tracking != null &&
|
|
trackingDirectivesEqual(expectedTracking, currentOp.tracking)
|
|
) {
|
|
// The current op has the right tracking. Extend it.
|
|
currentOp.r += retainedText
|
|
} else {
|
|
// The current op doesn't have the right tracking. Emit the current
|
|
// op and start a new one.
|
|
if (currentOp != null) {
|
|
ops.push(currentOp)
|
|
}
|
|
currentOp = {
|
|
r: retainedText,
|
|
p: cursor,
|
|
tracking: expectedTracking,
|
|
}
|
|
}
|
|
}
|
|
|
|
// Advance cursor
|
|
cursor = transition.pos
|
|
}
|
|
|
|
// Update the expected and persisted tracking
|
|
if (transition.stage === 'persisted') {
|
|
persistedTracking = transition.tracking
|
|
} else {
|
|
expectedTracking = transition.tracking
|
|
}
|
|
}
|
|
|
|
// Emit the last op
|
|
if (currentOp != null) {
|
|
ops.push(currentOp)
|
|
}
|
|
|
|
if (ops.length > 0) {
|
|
this.expandedUpdates.push({
|
|
doc: update.doc,
|
|
op: ops,
|
|
meta: {
|
|
resync: true,
|
|
origin: this.origin,
|
|
ts: update.meta.ts,
|
|
pathname,
|
|
doc_length: expectedContent.length,
|
|
},
|
|
})
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Compares the ranges in the persisted and expected comments
|
|
*
|
|
* @param {HistoryComment} persistedComment
|
|
* @param {Comment} expectedComment
|
|
*/
|
|
function commentRangesAreInSync(persistedComment, expectedComment) {
|
|
const expectedPos = expectedComment.op.hpos ?? expectedComment.op.p
|
|
const expectedLength = expectedComment.op.hlen ?? expectedComment.op.c.length
|
|
if (expectedLength === 0) {
|
|
// A zero length comment from RangesManager is a detached comment in history
|
|
return persistedComment.ranges.length === 0
|
|
}
|
|
|
|
if (persistedComment.ranges.length !== 1) {
|
|
// The editor only supports single range comments
|
|
return false
|
|
}
|
|
const persistedRange = persistedComment.ranges[0]
|
|
return (
|
|
persistedRange.pos === expectedPos &&
|
|
persistedRange.length === expectedLength
|
|
)
|
|
}
|
|
|
|
/**
|
|
* Iterates through expected tracked changes and persisted tracked changes and
|
|
* returns all transitions, sorted by position.
|
|
*
|
|
* @param {readonly HistoryTrackedChange[]} persistedChanges
|
|
* @param {TrackedChange[]} expectedChanges
|
|
* @param {number} docLength
|
|
*/
|
|
function getTrackedChangesTransitions(
|
|
persistedChanges,
|
|
expectedChanges,
|
|
docLength
|
|
) {
|
|
/** @type {TrackedChangeTransition[]} */
|
|
const transitions = []
|
|
|
|
for (const change of persistedChanges) {
|
|
transitions.push({
|
|
stage: 'persisted',
|
|
pos: change.range.start,
|
|
tracking: {
|
|
type: change.tracking.type,
|
|
userId: change.tracking.userId,
|
|
ts: change.tracking.ts.toISOString(),
|
|
},
|
|
})
|
|
transitions.push({
|
|
stage: 'persisted',
|
|
pos: change.range.end,
|
|
tracking: { type: 'none' },
|
|
})
|
|
}
|
|
|
|
for (const change of expectedChanges) {
|
|
const op = change.op
|
|
const pos = op.hpos ?? op.p
|
|
if (isInsert(op)) {
|
|
transitions.push({
|
|
stage: 'expected',
|
|
pos,
|
|
tracking: {
|
|
type: 'insert',
|
|
userId: change.metadata.user_id,
|
|
ts: change.metadata.ts,
|
|
},
|
|
})
|
|
transitions.push({
|
|
stage: 'expected',
|
|
pos: pos + op.i.length,
|
|
tracking: { type: 'none' },
|
|
})
|
|
} else {
|
|
transitions.push({
|
|
stage: 'expected',
|
|
pos,
|
|
tracking: {
|
|
type: 'delete',
|
|
userId: change.metadata.user_id,
|
|
ts: change.metadata.ts,
|
|
},
|
|
})
|
|
transitions.push({
|
|
stage: 'expected',
|
|
pos: pos + op.d.length,
|
|
tracking: { type: 'none' },
|
|
})
|
|
}
|
|
}
|
|
|
|
transitions.push({
|
|
stage: 'expected',
|
|
pos: docLength,
|
|
tracking: { type: 'none' },
|
|
})
|
|
|
|
transitions.sort((a, b) => {
|
|
if (a.pos < b.pos) {
|
|
return -1
|
|
} else if (a.pos > b.pos) {
|
|
return 1
|
|
} else if (a.tracking.type === 'none' && b.tracking.type !== 'none') {
|
|
// none type comes before other types so that it can be overridden at the
|
|
// same position
|
|
return -1
|
|
} else if (a.tracking.type !== 'none' && b.tracking.type === 'none') {
|
|
// none type comes before other types so that it can be overridden at the
|
|
// same position
|
|
return 1
|
|
} else {
|
|
return 0
|
|
}
|
|
})
|
|
|
|
return transitions
|
|
}
|
|
|
|
/**
|
|
* Returns true if both tracking directives are equal
|
|
*
|
|
* @param {TrackingDirective} a
|
|
* @param {TrackingDirective} b
|
|
*/
|
|
function trackingDirectivesEqual(a, b) {
|
|
if (a.type === 'none') {
|
|
return b.type === 'none'
|
|
} else {
|
|
return a.type === b.type && a.userId === b.userId && a.ts === b.ts
|
|
}
|
|
}
|
|
|
|
// EXPORTS
|
|
|
|
const startResyncCb = callbackify(startResync)
|
|
const startHardResyncCb = callbackify(startHardResync)
|
|
const setResyncStateCb = callbackify(setResyncState)
|
|
const clearResyncStateCb = callbackify(clearResyncState)
|
|
const skipUpdatesDuringSyncCb = callbackifyMultiResult(skipUpdatesDuringSync, [
|
|
'updates',
|
|
'syncState',
|
|
])
|
|
const expandSyncUpdatesCb = (
|
|
projectId,
|
|
projectHistoryId,
|
|
updates,
|
|
extendLock,
|
|
callback
|
|
) => {
|
|
const extendLockPromises = promisify(extendLock)
|
|
expandSyncUpdates(projectId, projectHistoryId, updates, extendLockPromises)
|
|
.then(result => {
|
|
callback(null, result)
|
|
})
|
|
.catch(err => {
|
|
callback(err)
|
|
})
|
|
}
|
|
|
|
export {
|
|
startResyncCb as startResync,
|
|
startHardResyncCb as startHardResync,
|
|
setResyncStateCb as setResyncState,
|
|
clearResyncStateCb as clearResyncState,
|
|
skipUpdatesDuringSyncCb as skipUpdatesDuringSync,
|
|
expandSyncUpdatesCb as expandSyncUpdates,
|
|
}
|
|
|
|
export const promises = {
|
|
startResync,
|
|
startHardResync,
|
|
setResyncState,
|
|
clearResyncState,
|
|
skipUpdatesDuringSync,
|
|
expandSyncUpdates,
|
|
}
|