mirror of
https://github.com/overleaf/overleaf.git
synced 2024-11-21 20:47:08 -05:00
26f3f3e2e2
Migrate scripts folder to esm 1/x GitOrigin-RevId: 4a4bc9a161f144fdb40ce3f2a0a9313b36c6df81
442 lines
13 KiB
JavaScript
442 lines
13 KiB
JavaScript
import minimist from 'minimist'
|
|
import settings from '@overleaf/settings'
|
|
import ProjectDetailsHandler from '../app/src/Features/Project/ProjectDetailsHandler.js'
|
|
import mongodb from '../app/src/infrastructure/mongodb.js'
|
|
import mongoose from '../app/src/infrastructure/Mongoose.js'
|
|
import fs from 'fs'
|
|
import path from 'path'
|
|
import crypto from 'crypto'
|
|
import fetch from 'node-fetch'
|
|
import http from 'http'
|
|
import _ from 'lodash'
|
|
|
|
const { ObjectId } = mongodb
|
|
|
|
// Examples:
|
|
//
|
|
// Simple usage:
|
|
// node stress_test.mjs --project-id=ID -n 100 --download-zip # download 100 zips from history-v1
|
|
// node stress_test.mjs --project-id=ID -n 100 --create-blob # create 100 blobs in history-v1
|
|
// node stress_test.mjs --project-id=ID -n 100 --fetch-blob # create blob and fetch it 100 times from history-v1
|
|
// node stress_test.mjs --project-id=ID -n 100 --upload-file # upload 100 files to filestore
|
|
// node stress_test.mjs --project-id=ID -n 100 --download-file # create file in filestore and download it 100 times
|
|
//
|
|
// Delay between requests:
|
|
// node stress_test.mjs --project-id=ID -n 100 --download-zip --sleep=0.1 # download 100 zips from history-v1 with 0.1s sleep
|
|
//
|
|
// Abort requests at random times:
|
|
// node stress_test.mjs --project-id=ID -n 100 --download-zip --abort # download 100 zips from history-v1 with aborts
|
|
//
|
|
// Parallel workers:
|
|
// node stress_test.mjs --project-id=ID -n 1000 -j 10 --upload-file # upload 1000 files in 10 parallel workers
|
|
//
|
|
// Fixed file size:
|
|
// node stress_test.mjs --project-id=ID -n 1000 --size 1000000 --upload-file # upload 1000 files of 1MB in 10 parallel workers
|
|
//
|
|
// Random file size:
|
|
// node stress_test.mjs --project-id=ID -n 1000 --size-min 1024 --size-max 10000000 --upload-file # upload 1000 files of 1KB to 10MB in 10 parallel workers
|
|
|
|
const argv = minimist(process.argv.slice(2), {
|
|
string: ['n', 'j', 'project-id', 'sleep', 'size', 'size-min', 'size-max'],
|
|
boolean: [
|
|
'download-zip',
|
|
'create-blob',
|
|
'fetch-blob',
|
|
'upload-file',
|
|
'download-file',
|
|
'use-file',
|
|
'abort',
|
|
],
|
|
default: {
|
|
n: 1,
|
|
j: 1,
|
|
sleep: 1,
|
|
size: 100 * 1024,
|
|
highWaterMark: 64 * 1024,
|
|
},
|
|
})
|
|
|
|
const projectId = argv['project-id']
|
|
if (!projectId) {
|
|
console.error(
|
|
'Usage: node stress_test.mjs --project-id ID -n COUNT -j CONCURRENCY --sleep T --size BYTES --use-file --[create-blob|fetch-blob|download-zip|upload-file|download-file]'
|
|
)
|
|
process.exit(1)
|
|
}
|
|
|
|
process.on('exit', () => {
|
|
log('Exiting')
|
|
})
|
|
|
|
async function sleep() {
|
|
const ms = argv.sleep * 1000 * (0.5 + Math.random())
|
|
return new Promise(resolve => setTimeout(resolve, ms))
|
|
}
|
|
function log(...args) {
|
|
const date = new Date()
|
|
console.log(date.toISOString(), ...args)
|
|
}
|
|
|
|
let abortTime = 1000
|
|
function adjustAbortTime(aborted, dt) {
|
|
if (!argv.abort) {
|
|
return
|
|
}
|
|
// If the last task was aborted, increase the abort time gradually
|
|
// Otherwise, reset the abort time to a random fraction of the response time.
|
|
if (aborted) {
|
|
abortTime = Math.min(abortTime * 1.5, 10000)
|
|
} else {
|
|
abortTime = Math.random() * dt
|
|
}
|
|
// Clamp to valid AbortSignal times
|
|
abortTime = Math.max(1, Math.round(abortTime))
|
|
}
|
|
|
|
function abortSignal() {
|
|
if (!argv.abort) {
|
|
return
|
|
}
|
|
return AbortSignal.timeout(abortTime)
|
|
}
|
|
|
|
async function stressTest(testCase, numberOfRuns, concurrentJobs) {
|
|
process.on('SIGINT', () => {
|
|
log('Caught interrupt signal. Running cleanup...')
|
|
numberOfRuns = 0
|
|
})
|
|
let startedTasks = 0
|
|
let finishedTasks = 0
|
|
let abortedTasks = 0
|
|
const periodicLog = _.throttle(log, 1000, { leading: true })
|
|
const errors = []
|
|
const { action, cleanup } = testCase
|
|
const executeTask = async () => {
|
|
startedTasks++
|
|
await sleep()
|
|
const t0 = Date.now()
|
|
try {
|
|
await action(abortSignal())
|
|
finishedTasks++
|
|
adjustAbortTime(false, Date.now() - t0)
|
|
} catch (err) {
|
|
if (err.name === 'AbortError') {
|
|
abortedTasks++
|
|
adjustAbortTime(true, Date.now() - t0)
|
|
} else {
|
|
errors.push(err)
|
|
log(startedTasks, err)
|
|
}
|
|
} finally {
|
|
periodicLog(
|
|
`Completed ${finishedTasks} / Aborted ${abortedTasks} / Errors ${errors.length}`
|
|
)
|
|
}
|
|
if (startedTasks < numberOfRuns) {
|
|
await executeTask()
|
|
}
|
|
}
|
|
const workers = []
|
|
for (let i = 0; i < concurrentJobs; i++) {
|
|
workers.push(executeTask())
|
|
}
|
|
|
|
try {
|
|
await Promise.all(workers)
|
|
periodicLog.cancel()
|
|
log(
|
|
`Completed ${finishedTasks} / Aborted ${abortedTasks} / Errors ${errors.length}`
|
|
)
|
|
log(startedTasks, 'tasks completed')
|
|
if (cleanup) {
|
|
log('Cleaning up')
|
|
try {
|
|
await cleanup()
|
|
} catch (err) {
|
|
log('error cleaning up', err)
|
|
}
|
|
}
|
|
} catch (err) {
|
|
log('error running stress test', err)
|
|
}
|
|
if (errors.length > 0) {
|
|
log('Errors:', errors.length)
|
|
throw new Error('Errors')
|
|
}
|
|
}
|
|
|
|
function generateRandomBuffer(size) {
|
|
if (argv['fill-string']) {
|
|
const buffer = Buffer.alloc(size, argv['fill-string'])
|
|
// add some randomness at the start to avoid every random buffer being the same
|
|
buffer.write(crypto.randomUUID())
|
|
return buffer
|
|
} else {
|
|
return Buffer.alloc(size, crypto.randomUUID())
|
|
}
|
|
}
|
|
|
|
function computeGitHash(buffer) {
|
|
const byteLength = buffer.byteLength
|
|
const hash = crypto.createHash('sha1')
|
|
hash.setEncoding('hex')
|
|
hash.update('blob ' + byteLength + '\x00')
|
|
hash.update(buffer)
|
|
hash.end()
|
|
return { hashHex: hash.read(), byteLength }
|
|
}
|
|
|
|
function computeMD5Hash(buffer) {
|
|
const hash = crypto.createHash('md5')
|
|
hash.update(buffer)
|
|
return hash.digest('hex')
|
|
}
|
|
|
|
function readableSize(size) {
|
|
// convert a size in bytes to a human readable string
|
|
const units = ['B', 'KB', 'MB', 'GB']
|
|
let i = 0
|
|
while (size > 1024 && i < units.length) {
|
|
size /= 1024
|
|
i++
|
|
}
|
|
return `${size.toFixed(2)} ${units[i]}`.trim()
|
|
}
|
|
|
|
class SizeGenerator {
|
|
constructor() {
|
|
if (argv['size-min'] && argv['size-max']) {
|
|
this.size_min = parseInt(argv['size-min']) || 0
|
|
this.size_max = parseInt(argv['size-max']) || argv.size
|
|
log(
|
|
`File size range [${readableSize(this.size_min)}, ${readableSize(
|
|
this.size_max
|
|
)}]`
|
|
)
|
|
} else {
|
|
this.size = parseInt(argv.size)
|
|
this.fixed = true
|
|
log('File size', readableSize(this.size))
|
|
}
|
|
}
|
|
|
|
get() {
|
|
return this.fixed
|
|
? this.size
|
|
: this.size_min + Math.random() * (this.size_max - this.size_min)
|
|
}
|
|
}
|
|
|
|
async function createBlob(projectId) {
|
|
log('Getting history id')
|
|
const v1Id = await getHistoryId(projectId)
|
|
// generate a random blob in a buffer and compute the git hash of the buffer
|
|
log('Creating test blob')
|
|
const userSize = new SizeGenerator()
|
|
async function putBlob(abortSignal) {
|
|
// create a random buffer and compute its hash
|
|
const buffer = generateRandomBuffer(userSize.get())
|
|
const { hashHex, byteLength } = computeGitHash(buffer)
|
|
// write the buffer to a file for streaming
|
|
let readStream
|
|
let filepath
|
|
if (argv['use-file']) {
|
|
filepath = path.join('/tmp', `${v1Id}-${hashHex}-${crypto.randomUUID()}`)
|
|
await fs.promises.writeFile(filepath, buffer)
|
|
const filestream = fs.createReadStream(filepath, {
|
|
highWaterMark: argv.highWaterMark,
|
|
})
|
|
readStream = filestream
|
|
} else {
|
|
filepath = null
|
|
readStream = buffer
|
|
}
|
|
const putUrl = `${settings.apis.v1_history.url}/projects/${v1Id}/blobs/${hashHex}`
|
|
const options = {
|
|
method: 'PUT',
|
|
headers: {
|
|
'Content-Type': 'application/octet-stream',
|
|
'Content-Length': byteLength,
|
|
Authorization: `Basic ${Buffer.from(
|
|
`${settings.apis.v1_history.user}:${settings.apis.v1_history.pass}`
|
|
).toString('base64')}`,
|
|
},
|
|
}
|
|
const req = http.request(putUrl, options)
|
|
return await new Promise((resolve, reject) => {
|
|
req.on('error', reject)
|
|
req.on('response', res => {
|
|
if (res.statusCode !== 201) {
|
|
reject(
|
|
new Error(
|
|
`failed to put blob ${putUrl} status=${res.statusCode} ${res.statusMessage}`
|
|
)
|
|
)
|
|
} else {
|
|
resolve({ hashHex, byteLength })
|
|
}
|
|
})
|
|
readStream.pipe(req)
|
|
})
|
|
}
|
|
return { action: putBlob, description: 'createBlob in history-v1' }
|
|
}
|
|
|
|
async function fetchBlob(projectId) {
|
|
log('Getting history id and creating test blob')
|
|
const v1Id = await getHistoryId(projectId)
|
|
const { action: putBlob } = await createBlob(projectId)
|
|
const { hashHex, byteLength } = await putBlob()
|
|
async function getBlob(abortSignal) {
|
|
const getUrl = `${settings.apis.v1_history.url}/projects/${v1Id}/blobs/${hashHex}`
|
|
const response = await historyFetch(getUrl, { signal: abortSignal })
|
|
if (!response.ok) {
|
|
throw new Error(`failed to get blob ${getUrl} status=${response.status}`)
|
|
}
|
|
const buffer = await response.arrayBuffer()
|
|
if (buffer.byteLength !== byteLength) {
|
|
throw new Error(
|
|
`unexpected fetch blob length ${buffer.byteLength} vs expected ${byteLength}`
|
|
)
|
|
}
|
|
}
|
|
return { action: getBlob, description: 'fetchBlob from history-v1' }
|
|
}
|
|
|
|
async function getHistoryId(projectId) {
|
|
const project = await ProjectDetailsHandler.promises.getDetails(projectId)
|
|
const v1Id = project?.overleaf?.history?.id
|
|
return v1Id
|
|
}
|
|
|
|
async function downloadHistoryZip(projectId) {
|
|
log('Getting history id and latest version')
|
|
const v1Id = await getHistoryId(projectId)
|
|
const latestUrl = `${settings.apis.v1_history.url}/projects/${v1Id}/latest/history`
|
|
let response = await historyFetch(latestUrl)
|
|
if (!response.ok) {
|
|
throw new Error(
|
|
`failed to get latest version ${latestUrl} status=${response.status}`
|
|
)
|
|
}
|
|
const latestBody = await response.json()
|
|
const version =
|
|
latestBody.chunk.startVersion + latestBody.chunk.history.changes.length
|
|
const zipUrl = `${settings.apis.v1_history.url}/projects/${v1Id}/version/${version}/zip`
|
|
let expectedLength = null
|
|
async function getZip(abortSignal) {
|
|
response = await historyFetch(zipUrl, { signal: abortSignal })
|
|
const responseBuffer = await response.arrayBuffer()
|
|
if (expectedLength === null) {
|
|
expectedLength = responseBuffer.byteLength
|
|
} else if (responseBuffer.byteLength !== expectedLength) {
|
|
throw new Error(
|
|
`unexpected zip download length ${responseBuffer.byteLength} vs expected ${expectedLength}`
|
|
)
|
|
}
|
|
}
|
|
return { action: getZip, description: 'download zip from history-v1' }
|
|
}
|
|
|
|
async function historyFetch(url, options) {
|
|
const authHeader = {
|
|
Authorization: `Basic ${Buffer.from(
|
|
`${settings.apis.v1_history.user}:${settings.apis.v1_history.pass}`
|
|
).toString('base64')}`,
|
|
}
|
|
const response = await fetch(url, { ...options, headers: authHeader })
|
|
if (!response.ok) {
|
|
throw new Error(`failed to download url ${url} status=${response.status}`)
|
|
}
|
|
return response
|
|
}
|
|
|
|
async function _deleteFile(url, log) {
|
|
const response = await fetch(url, { method: 'DELETE' })
|
|
if (!response.ok) {
|
|
throw new Error(`failed to delete file status=${response.status}`)
|
|
}
|
|
}
|
|
|
|
async function uploadFile(projectId) {
|
|
// generate a random blob in a buffer and compute the md5 hash of the buffer
|
|
const userSize = new SizeGenerator()
|
|
async function upload(abortSignal, deleteFile = true) {
|
|
const size = userSize.get()
|
|
const buffer = Buffer.alloc(size, crypto.randomUUID())
|
|
const fileId = new ObjectId()
|
|
const url = `${settings.apis.filestore.url}/project/${projectId}/file/${fileId}`
|
|
const md5 = computeMD5Hash(buffer)
|
|
const response = await fetch(url, {
|
|
method: 'POST',
|
|
body: buffer,
|
|
signal: abortSignal,
|
|
})
|
|
if (!response.ok) {
|
|
throw new Error(`failed to upload file ${url} status=${response.status}`)
|
|
}
|
|
if (deleteFile) {
|
|
await _deleteFile(url)
|
|
}
|
|
return { url, md5 }
|
|
}
|
|
return { action: upload, description: 'upload file to filestore' }
|
|
}
|
|
|
|
async function downloadFile(projectId) {
|
|
log('Creating test file')
|
|
const { action: upload } = await uploadFile(projectId)
|
|
const { url, md5: expectedMd5 } = await upload(null, false)
|
|
async function download(abortSignal) {
|
|
const response = await fetch(url, {
|
|
method: 'GET',
|
|
signal: abortSignal,
|
|
})
|
|
if (!response.ok) {
|
|
throw new Error(`failed to get file ${url} status=${response.status}`)
|
|
}
|
|
const md5 = computeMD5Hash(Buffer.from(await response.arrayBuffer()))
|
|
if (md5 !== expectedMd5) {
|
|
throw new Error(`md5 mismatch`)
|
|
}
|
|
}
|
|
async function cleanup() {
|
|
log('Deleting test file')
|
|
await _deleteFile(url)
|
|
}
|
|
return {
|
|
action: download,
|
|
cleanup,
|
|
description: 'download file from filestore',
|
|
}
|
|
}
|
|
|
|
async function run() {
|
|
let testCase
|
|
if (argv['download-zip']) {
|
|
testCase = await downloadHistoryZip(projectId)
|
|
} else if (argv['create-blob']) {
|
|
testCase = await createBlob(projectId)
|
|
} else if (argv['fetch-blob']) {
|
|
testCase = await fetchBlob(projectId)
|
|
} else if (argv['upload-file']) {
|
|
testCase = await uploadFile(projectId)
|
|
} else if (argv['download-file']) {
|
|
testCase = await downloadFile(projectId)
|
|
} else {
|
|
throw new Error('unknown command')
|
|
}
|
|
log('Running stress test:', testCase.description)
|
|
await stressTest(testCase, argv.n, argv.j)
|
|
log('Stress test done')
|
|
}
|
|
|
|
try {
|
|
await Promise.all([mongodb.waitForDb(), mongoose.connectionPromise])
|
|
await run()
|
|
log('Completed')
|
|
process.exit(0)
|
|
} catch (error) {
|
|
console.error(error)
|
|
process.exit(1)
|
|
}
|