| | |
| | | import { Processor, unified } from "unified" |
| | | import { Root as MDRoot } from "remark-parse/lib" |
| | | import { Root as HTMLRoot } from "hast" |
| | | import { ProcessedContent } from "../plugins/vfile" |
| | | import { MarkdownContent, ProcessedContent } from "../plugins/vfile" |
| | | import { PerfTimer } from "../util/perf" |
| | | import { read } from "to-vfile" |
| | | import { FilePath, QUARTZ, slugifyFilePath } from "../util/path" |
| | |
| | | import workerpool, { Promise as WorkerPromise } from "workerpool" |
| | | import { QuartzLogger } from "../util/log" |
| | | import { trace } from "../util/trace" |
| | | import { BuildCtx } from "../util/ctx" |
| | | import { BuildCtx, WorkerSerializableBuildCtx } from "../util/ctx" |
| | | import { styleText } from "util" |
| | | |
| | | export type QuartzProcessor = Processor<MDRoot, MDRoot, HTMLRoot> |
| | | export function createProcessor(ctx: BuildCtx): QuartzProcessor { |
| | | export type QuartzMdProcessor = Processor<MDRoot, MDRoot, MDRoot> |
| | | export type QuartzHtmlProcessor = Processor<undefined, MDRoot, HTMLRoot> |
| | | |
| | | export function createMdProcessor(ctx: BuildCtx): QuartzMdProcessor { |
| | | const transformers = ctx.cfg.plugins.transformers |
| | | |
| | | return ( |
| | |
| | | .use(remarkParse) |
| | | // MD AST -> MD AST transforms |
| | | .use( |
| | | transformers |
| | | .filter((p) => p.markdownPlugins) |
| | | .flatMap((plugin) => plugin.markdownPlugins!(ctx)), |
| | | ) |
| | | transformers.flatMap((plugin) => plugin.markdownPlugins?.(ctx) ?? []), |
| | | ) as unknown as QuartzMdProcessor |
| | | // ^ sadly the typing of `use` is not smart enough to infer the correct type from our plugin list |
| | | ) |
| | | } |
| | | |
| | | export function createHtmlProcessor(ctx: BuildCtx): QuartzHtmlProcessor { |
| | | const transformers = ctx.cfg.plugins.transformers |
| | | return ( |
| | | unified() |
| | | // MD AST -> HTML AST |
| | | .use(remarkRehype, { allowDangerousHtml: true }) |
| | | // HTML AST -> HTML AST transforms |
| | | .use(transformers.filter((p) => p.htmlPlugins).flatMap((plugin) => plugin.htmlPlugins!(ctx))) |
| | | .use(transformers.flatMap((plugin) => plugin.htmlPlugins?.(ctx) ?? [])) |
| | | ) |
| | | } |
| | | |
| | |
| | | |
| | | export function createFileParser(ctx: BuildCtx, fps: FilePath[]) { |
| | | const { argv, cfg } = ctx |
| | | return async (processor: QuartzProcessor) => { |
| | | const res: ProcessedContent[] = [] |
| | | return async (processor: QuartzMdProcessor) => { |
| | | const res: MarkdownContent[] = [] |
| | | for (const fp of fps) { |
| | | try { |
| | | const perf = new PerfTimer() |
| | |
| | | res.push([newAst, file]) |
| | | |
| | | if (argv.verbose) { |
| | | console.log(`[process] ${fp} -> ${file.data.slug} (${perf.timeSince()})`) |
| | | console.log(`[markdown] ${fp} -> ${file.data.slug} (${perf.timeSince()})`) |
| | | } |
| | | } catch (err) { |
| | | trace(`\nFailed to process \`${fp}\``, err as Error) |
| | | trace(`\nFailed to process markdown \`${fp}\``, err as Error) |
| | | } |
| | | } |
| | | |
| | | return res |
| | | } |
| | | } |
| | | |
| | | export function createMarkdownParser(ctx: BuildCtx, mdContent: MarkdownContent[]) { |
| | | return async (processor: QuartzHtmlProcessor) => { |
| | | const res: ProcessedContent[] = [] |
| | | for (const [ast, file] of mdContent) { |
| | | try { |
| | | const perf = new PerfTimer() |
| | | |
| | | const newAst = await processor.run(ast as MDRoot, file) |
| | | res.push([newAst, file]) |
| | | |
| | | if (ctx.argv.verbose) { |
| | | console.log(`[html] ${file.data.slug} (${perf.timeSince()})`) |
| | | } |
| | | } catch (err) { |
| | | trace(`\nFailed to process html \`${file.data.filePath}\``, err as Error) |
| | | } |
| | | } |
| | | |
| | |
| | | |
| | | const clamp = (num: number, min: number, max: number) => |
| | | Math.min(Math.max(Math.round(num), min), max) |
| | | |
| | | export async function parseMarkdown(ctx: BuildCtx, fps: FilePath[]): Promise<ProcessedContent[]> { |
| | | const { argv } = ctx |
| | | const perf = new PerfTimer() |
| | |
| | | log.start(`Parsing input files using ${concurrency} threads`) |
| | | if (concurrency === 1) { |
| | | try { |
| | | const processor = createProcessor(ctx) |
| | | const parse = createFileParser(ctx, fps) |
| | | res = await parse(processor) |
| | | const mdRes = await createFileParser(ctx, fps)(createMdProcessor(ctx)) |
| | | res = await createMarkdownParser(ctx, mdRes)(createHtmlProcessor(ctx)) |
| | | } catch (error) { |
| | | log.end() |
| | | throw error |
| | |
| | | maxWorkers: concurrency, |
| | | workerType: "thread", |
| | | }) |
| | | |
| | | const childPromises: WorkerPromise<ProcessedContent[]>[] = [] |
| | | for (const chunk of chunks(fps, CHUNK_SIZE)) { |
| | | childPromises.push(pool.exec("parseFiles", [argv, chunk, ctx.allSlugs])) |
| | | const errorHandler = (err: any) => { |
| | | console.error(err) |
| | | process.exit(1) |
| | | } |
| | | |
| | | const results: ProcessedContent[][] = await WorkerPromise.all(childPromises).catch((err) => { |
| | | const errString = err.toString().slice("Error:".length) |
| | | console.error(errString) |
| | | process.exit(1) |
| | | }) |
| | | const serializableCtx: WorkerSerializableBuildCtx = { |
| | | buildId: ctx.buildId, |
| | | argv: ctx.argv, |
| | | allSlugs: ctx.allSlugs, |
| | | allFiles: ctx.allFiles, |
| | | incremental: ctx.incremental, |
| | | } |
| | | |
| | | const textToMarkdownPromises: WorkerPromise<MarkdownContent[]>[] = [] |
| | | let processedFiles = 0 |
| | | for (const chunk of chunks(fps, CHUNK_SIZE)) { |
| | | textToMarkdownPromises.push(pool.exec("parseMarkdown", [serializableCtx, chunk])) |
| | | } |
| | | |
| | | const mdResults: Array<MarkdownContent[]> = await Promise.all( |
| | | textToMarkdownPromises.map(async (promise) => { |
| | | const result = await promise |
| | | processedFiles += result.length |
| | | log.updateText(`text->markdown ${styleText("gray", `${processedFiles}/${fps.length}`)}`) |
| | | return result |
| | | }), |
| | | ).catch(errorHandler) |
| | | |
| | | const markdownToHtmlPromises: WorkerPromise<ProcessedContent[]>[] = [] |
| | | processedFiles = 0 |
| | | for (const mdChunk of mdResults) { |
| | | markdownToHtmlPromises.push(pool.exec("processHtml", [serializableCtx, mdChunk])) |
| | | } |
| | | const results: ProcessedContent[][] = await Promise.all( |
| | | markdownToHtmlPromises.map(async (promise) => { |
| | | const result = await promise |
| | | processedFiles += result.length |
| | | log.updateText(`markdown->html ${styleText("gray", `${processedFiles}/${fps.length}`)}`) |
| | | return result |
| | | }), |
| | | ).catch(errorHandler) |
| | | |
| | | res = results.flat() |
| | | await pool.terminate() |
| | | } |