profiling, better concurrency heuristics

This commit is contained in:
Jacky Zhao 2023-08-09 00:26:33 -07:00
parent 68ccd1d79d
commit cea6834fef
4 changed files with 9 additions and 6 deletions
quartz/processors

View file

@ -9,7 +9,6 @@ import { PerfTimer } from "../perf"
import { read } from "to-vfile"
import { FilePath, QUARTZ, slugifyFilePath } from "../path"
import path from "path"
import os from "os"
import workerpool, { Promise as WorkerPromise } from "workerpool"
import { QuartzLogger } from "../log"
import { trace } from "../trace"
@ -82,6 +81,7 @@ export function createFileParser(ctx: BuildCtx, fps: FilePath[]) {
const res: ProcessedContent[] = []
for (const fp of fps) {
try {
const perf = new PerfTimer()
const file = await read(fp)
// strip leading and trailing whitespace
@ -101,7 +101,7 @@ export function createFileParser(ctx: BuildCtx, fps: FilePath[]) {
res.push([newAst, file])
if (argv.verbose) {
console.log(`[process] ${fp} -> ${file.data.slug}`)
console.log(`[process] ${fp} -> ${file.data.slug} (${perf.timeSince()})`)
}
} catch (err) {
trace(`\nFailed to process \`${fp}\``, err as Error)
@ -112,14 +112,15 @@ export function createFileParser(ctx: BuildCtx, fps: FilePath[]) {
}
}
const clamp = (num: number, min: number, max: number) => Math.min(Math.max(Math.round(num), min), max);
export async function parseMarkdown(ctx: BuildCtx, fps: FilePath[]): Promise<ProcessedContent[]> {
const { argv } = ctx
const perf = new PerfTimer()
const log = new QuartzLogger(argv.verbose)
// rough heuristics: 128 gives enough time for v8 to JIT and optimize parsing code paths
const CHUNK_SIZE = 128
let concurrency =
ctx.argv.concurrency ?? (fps.length < CHUNK_SIZE ? 1 : os.availableParallelism())
const concurrency = ctx.argv.concurrency ?? clamp(fps.length / CHUNK_SIZE, 1, 4)
let res: ProcessedContent[] = []
log.start(`Parsing input files using ${concurrency} threads`)