worm-scraper/lib/convert.js
2021-02-15 22:05:38 -05:00

48 lines
1.7 KiB
JavaScript

"use strict";
const path = require("path");
const fs = require("fs").promises;
const { performance } = require("perf_hooks");
const workerpool = require("workerpool");
const cliProgress = require("cli-progress");
module.exports = async (cachePath, manifestPath, contentPath, book, concurrentJobs) => {
const manifestContents = await fs.readFile(manifestPath, { encoding: "utf-8" });
const chapters = JSON.parse(manifestContents);
console.log("Converting raw downloaded HTML to EPUB chapters");
const progress = new cliProgress.SingleBar({
stopOnComplete: true,
clearOnComplete: true,
format: " {bar} {percentage}% | {time} | {value}/{total}"
}, cliProgress.Presets.shades_classic);
const start = performance.now();
progress.start(chapters.length, 0, { time: " " });
const poolOptions = {};
if (concurrentJobs !== undefined) {
poolOptions.maxWorkers = concurrentJobs;
}
const pool = workerpool.pool(path.resolve(__dirname, "convert-worker.js"), poolOptions);
const warnings = [];
await Promise.all(chapters.map(async chapter => {
const inputPath = path.resolve(cachePath, chapter.filename);
const destFileName = `${path.basename(chapter.filename, ".html")}.xhtml`;
const outputPath = path.resolve(contentPath, destFileName);
warnings.push(...await pool.exec("convertChapter", [chapter, book, inputPath, outputPath]));
const seconds = String(Math.round((performance.now() - start) / 1000)).padStart(3);
progress.increment({ time: `${seconds} s` });
}));
pool.terminate();
for (const warning of warnings) {
console.warn(warning);
}
console.log(`All chapters converted in ${Math.round((performance.now() - start) / 100) / 10} seconds`);
};