import sourceMapSupport from "source-map-support" sourceMapSupport.install(options) import path from "path" import { PerfTimer } from "./util/perf" import { rimraf } from "rimraf" import { GlobbyFilterFunction, isGitIgnored } from "globby" import chalk from "chalk" import { parseMarkdown } from "./processors/parse" import { filterContent } from "./processors/filter" import { emitContent } from "./processors/emit" import cfg from "../quartz.config" import { FilePath, FullSlug, joinSegments, slugifyFilePath } from "./util/path" import chokidar from "chokidar" import { ProcessedContent } from "./plugins/vfile" import { Argv, BuildCtx } from "./util/ctx" import { glob, toPosixPath } from "./util/glob" import { trace } from "./util/trace" import { options } from "./util/sourcemap" import { Mutex } from "async-mutex" import DepGraph from "./depgraph" import { getStaticResourcesFromPlugins } from "./plugins" type Dependencies = Record | null> type BuildData = { ctx: BuildCtx ignored: GlobbyFilterFunction mut: Mutex initialSlugs: FullSlug[] // TODO merge contentMap and trackedAssets contentMap: Map trackedAssets: Set toRebuild: Set toRemove: Set lastBuildMs: number dependencies: Dependencies } type FileEvent = "add" | "change" | "delete" function newBuildId() { return Math.random().toString(36).substring(2, 8) } async function buildQuartz(argv: Argv, mut: Mutex, clientRefresh: () => void) { const ctx: BuildCtx = { buildId: newBuildId(), argv, cfg, allSlugs: [], } const perf = new PerfTimer() const output = argv.output const pluginCount = Object.values(cfg.plugins).flat().length const pluginNames = (key: "transformers" | "filters" | "emitters") => cfg.plugins[key].map((plugin) => plugin.name) if (argv.verbose) { console.log(`Loaded ${pluginCount} plugins`) console.log(` Transformers: ${pluginNames("transformers").join(", ")}`) console.log(` Filters: ${pluginNames("filters").join(", ")}`) console.log(` Emitters: ${pluginNames("emitters").join(", ")}`) } const release = await mut.acquire() perf.addEvent("clean") await rimraf(path.join(output, "*"), { glob: true }) console.log(`Cleaned output directory \`${output}\` in ${perf.timeSince("clean")}`) perf.addEvent("glob") const allFiles = await glob("**/*.*", argv.directory, cfg.configuration.ignorePatterns) const fps = allFiles.filter((fp) => fp.endsWith(".md")).sort() console.log( `Found ${fps.length} input files from \`${argv.directory}\` in ${perf.timeSince("glob")}`, ) const filePaths = fps.map((fp) => joinSegments(argv.directory, fp) as FilePath) ctx.allSlugs = allFiles.map((fp) => slugifyFilePath(fp as FilePath)) const parsedFiles = await parseMarkdown(ctx, filePaths) const filteredContent = filterContent(ctx, parsedFiles) const dependencies: Record | null> = {} // Only build dependency graphs if we're doing a fast rebuild if (argv.fastRebuild) { const staticResources = getStaticResourcesFromPlugins(ctx) for (const emitter of cfg.plugins.emitters) { dependencies[emitter.name] = (await emitter.getDependencyGraph?.(ctx, filteredContent, staticResources)) ?? null } } await emitContent(ctx, filteredContent) console.log(chalk.green(`Done processing ${fps.length} files in ${perf.timeSince()}`)) release() if (argv.serve) { return startServing(ctx, mut, parsedFiles, clientRefresh, dependencies) } } // setup watcher for rebuilds async function startServing( ctx: BuildCtx, mut: Mutex, initialContent: ProcessedContent[], clientRefresh: () => void, dependencies: Dependencies, // emitter name: dep graph ) { const { argv } = ctx // cache file parse results const contentMap = new Map() for (const content of initialContent) { const [_tree, vfile] = content contentMap.set(vfile.data.filePath!, content) } const buildData: BuildData = { ctx, mut, dependencies, contentMap, ignored: await isGitIgnored(), initialSlugs: ctx.allSlugs, toRebuild: new Set(), toRemove: new Set(), trackedAssets: new Set(), lastBuildMs: 0, } const watcher = chokidar.watch(".", { persistent: true, cwd: argv.directory, ignoreInitial: true, }) const buildFromEntry = argv.fastRebuild ? partialRebuildFromEntrypoint : rebuildFromEntrypoint watcher .on("add", (fp) => buildFromEntry(fp, "add", clientRefresh, buildData)) .on("change", (fp) => buildFromEntry(fp, "change", clientRefresh, buildData)) .on("unlink", (fp) => buildFromEntry(fp, "delete", clientRefresh, buildData)) return async () => { await watcher.close() } } async function partialRebuildFromEntrypoint( filepath: string, action: FileEvent, clientRefresh: () => void, buildData: BuildData, // note: this function mutates buildData ) { const { ctx, ignored, dependencies, contentMap, mut, toRemove } = buildData const { argv, cfg } = ctx // don't do anything for gitignored files if (ignored(filepath)) { return } const buildId = newBuildId() ctx.buildId = buildId buildData.lastBuildMs = new Date().getTime() const release = await mut.acquire() // if there's another build after us, release and let them do it if (ctx.buildId !== buildId) { release() return } const perf = new PerfTimer() console.log(chalk.yellow("Detected change, rebuilding...")) // UPDATE DEP GRAPH const fp = joinSegments(argv.directory, toPosixPath(filepath)) as FilePath const staticResources = getStaticResourcesFromPlugins(ctx) let processedFiles: ProcessedContent[] = [] switch (action) { case "add": // add to cache when new file is added processedFiles = await parseMarkdown(ctx, [fp]) processedFiles.forEach(([tree, vfile]) => contentMap.set(vfile.data.filePath!, [tree, vfile])) // update the dep graph by asking all emitters whether they depend on this file for (const emitter of cfg.plugins.emitters) { const emitterGraph = (await emitter.getDependencyGraph?.(ctx, processedFiles, staticResources)) ?? null if (emitterGraph) { const existingGraph = dependencies[emitter.name] if (existingGraph !== null) { existingGraph.mergeGraph(emitterGraph) } else { // might be the first time we're adding a mardown file dependencies[emitter.name] = emitterGraph } } } break case "change": // invalidate cache when file is changed processedFiles = await parseMarkdown(ctx, [fp]) processedFiles.forEach(([tree, vfile]) => contentMap.set(vfile.data.filePath!, [tree, vfile])) // only content files can have added/removed dependencies because of transclusions if (path.extname(fp) === ".md") { for (const emitter of cfg.plugins.emitters) { // get new dependencies from all emitters for this file const emitterGraph = (await emitter.getDependencyGraph?.(ctx, processedFiles, staticResources)) ?? null // only update the graph if the emitter plugin uses the changed file // eg. Assets plugin ignores md files, so we skip updating the graph if (emitterGraph?.hasNode(fp)) { // merge the new dependencies into the dep graph dependencies[emitter.name]?.updateIncomingEdgesForNode(emitterGraph, fp) } } } break case "delete": toRemove.add(fp) break } if (argv.verbose) { console.log(`Updated dependency graphs in ${perf.timeSince()}`) } // EMIT perf.addEvent("rebuild") let emittedFiles = 0 for (const emitter of cfg.plugins.emitters) { const depGraph = dependencies[emitter.name] // emitter hasn't defined a dependency graph. call it with all processed files if (depGraph === null) { if (argv.verbose) { console.log( `Emitter ${emitter.name} doesn't define a dependency graph. Calling it with all files...`, ) } const files = [...contentMap.values()].filter( ([_node, vfile]) => !toRemove.has(vfile.data.filePath!), ) const emittedFps = await emitter.emit(ctx, files, staticResources) if (ctx.argv.verbose) { for (const file of emittedFps) { console.log(`[emit:${emitter.name}] ${file}`) } } emittedFiles += emittedFps.length continue } // only call the emitter if it uses this file if (depGraph.hasNode(fp)) { // re-emit using all files that are needed for the downstream of this file // eg. for ContentIndex, the dep graph could be: // a.md --> contentIndex.json // b.md ------^ // // if a.md changes, we need to re-emit contentIndex.json, // and supply [a.md, b.md] to the emitter const upstreams = [...depGraph.getLeafNodeAncestors(fp)] as FilePath[] const upstreamContent = upstreams // filter out non-markdown files .filter((file) => contentMap.has(file)) // if file was deleted, don't give it to the emitter .filter((file) => !toRemove.has(file)) .map((file) => contentMap.get(file)!) const emittedFps = await emitter.emit(ctx, upstreamContent, staticResources) if (ctx.argv.verbose) { for (const file of emittedFps) { console.log(`[emit:${emitter.name}] ${file}`) } } emittedFiles += emittedFps.length } } console.log(`Emitted ${emittedFiles} files to \`${argv.output}\` in ${perf.timeSince("rebuild")}`) // CLEANUP const destinationsToDelete = new Set() for (const file of toRemove) { // remove from cache contentMap.delete(file) Object.values(dependencies).forEach((depGraph) => { // remove the node from dependency graphs depGraph?.removeNode(file) // remove any orphan nodes. eg if a.md is deleted, a.html is orphaned and should be removed const orphanNodes = depGraph?.removeOrphanNodes() orphanNodes?.forEach((node) => { // only delete files that are in the output directory if (node.startsWith(argv.output)) { destinationsToDelete.add(node) } }) }) } await rimraf([...destinationsToDelete]) console.log(chalk.green(`Done rebuilding in ${perf.timeSince()}`)) toRemove.clear() release() clientRefresh() } async function rebuildFromEntrypoint( fp: string, action: FileEvent, clientRefresh: () => void, buildData: BuildData, // note: this function mutates buildData ) { const { ctx, ignored, mut, initialSlugs, contentMap, toRebuild, toRemove, trackedAssets } = buildData const { argv } = ctx // don't do anything for gitignored files if (ignored(fp)) { return } // dont bother rebuilding for non-content files, just track and refresh fp = toPosixPath(fp) const filePath = joinSegments(argv.directory, fp) as FilePath if (path.extname(fp) !== ".md") { if (action === "add" || action === "change") { trackedAssets.add(filePath) } else if (action === "delete") { trackedAssets.delete(filePath) } clientRefresh() return } if (action === "add" || action === "change") { toRebuild.add(filePath) } else if (action === "delete") { toRemove.add(filePath) } const buildId = newBuildId() ctx.buildId = buildId buildData.lastBuildMs = new Date().getTime() const release = await mut.acquire() // there's another build after us, release and let them do it if (ctx.buildId !== buildId) { release() return } const perf = new PerfTimer() console.log(chalk.yellow("Detected change, rebuilding...")) try { const filesToRebuild = [...toRebuild].filter((fp) => !toRemove.has(fp)) const parsedContent = await parseMarkdown(ctx, filesToRebuild) for (const content of parsedContent) { const [_tree, vfile] = content contentMap.set(vfile.data.filePath!, content) } for (const fp of toRemove) { contentMap.delete(fp) } const parsedFiles = [...contentMap.values()] const filteredContent = filterContent(ctx, parsedFiles) // re-update slugs const trackedSlugs = [...new Set([...contentMap.keys(), ...toRebuild, ...trackedAssets])] .filter((fp) => !toRemove.has(fp)) .map((fp) => slugifyFilePath(path.posix.relative(argv.directory, fp) as FilePath)) ctx.allSlugs = [...new Set([...initialSlugs, ...trackedSlugs])] // TODO: we can probably traverse the link graph to figure out what's safe to delete here // instead of just deleting everything await rimraf(path.join(argv.output, ".*"), { glob: true }) await emitContent(ctx, filteredContent) console.log(chalk.green(`Done rebuilding in ${perf.timeSince()}`)) } catch (err) { console.log(chalk.yellow(`Rebuild failed. Waiting on a change to fix the error...`)) if (argv.verbose) { console.log(chalk.red(err)) } } clientRefresh() toRebuild.clear() toRemove.clear() release() } export default async (argv: Argv, mut: Mutex, clientRefresh: () => void) => { try { return await buildQuartz(argv, mut, clientRefresh) } catch (err) { trace("\nExiting Quartz due to a fatal error", err as Error) } }