Run add_omni_file script when file change detected
This commit is contained in:
parent
7aa3bde170
commit
8b76bf6e65
7 changed files with 86 additions and 34 deletions
|
@ -11,7 +11,7 @@ const ROOT = path.resolve(__dirname, '..');
|
|||
|
||||
async function getBrowserify(signatures) {
|
||||
const t1 = Date.now();
|
||||
var count = 0;
|
||||
const outFiles = [];
|
||||
var config, f, totalCount;
|
||||
|
||||
while ((config = browserifyConfigs.pop()) != null) {
|
||||
|
@ -48,7 +48,7 @@ async function getBrowserify(signatures) {
|
|||
|
||||
onProgress(f, dest, 'browserify');
|
||||
signatures[f] = newFileSignature;
|
||||
count++;
|
||||
outFiles.push(dest);
|
||||
} catch (err) {
|
||||
throw new Error(`Failed on ${f}: ${err}`);
|
||||
}
|
||||
|
@ -58,7 +58,8 @@ async function getBrowserify(signatures) {
|
|||
const t2 = Date.now();
|
||||
return {
|
||||
action: 'browserify',
|
||||
count,
|
||||
count: outFiles.length,
|
||||
outFiles,
|
||||
totalCount,
|
||||
processingTime: t2 - t1
|
||||
};
|
||||
|
|
|
@ -12,7 +12,7 @@ async function getCopy(source, options, signatures) {
|
|||
const t1 = Date.now();
|
||||
const files = await globby(source, Object.assign({ cwd: ROOT }, options ));
|
||||
const totalCount = files.length;
|
||||
var count = 0;
|
||||
const outFiles = [];
|
||||
var f;
|
||||
|
||||
while ((f = files.pop()) != null) {
|
||||
|
@ -34,7 +34,7 @@ async function getCopy(source, options, signatures) {
|
|||
await fs.copy(f, dest);
|
||||
onProgress(f, dest, 'cp');
|
||||
signatures[f] = newFileSignature;
|
||||
count++;
|
||||
outFiles.push(dest);
|
||||
} catch (err) {
|
||||
throw new Error(`Failed on ${f}: ${err}`);
|
||||
}
|
||||
|
@ -43,7 +43,8 @@ async function getCopy(source, options, signatures) {
|
|||
const t2 = Date.now();
|
||||
return {
|
||||
action: 'copy',
|
||||
count,
|
||||
count: outFiles.length,
|
||||
outFiles,
|
||||
totalCount,
|
||||
processingTime: t2 - t1
|
||||
};
|
||||
|
|
|
@ -14,7 +14,6 @@ async function getJS(source, options, signatures) {
|
|||
const matchingJSFiles = await globby(source, Object.assign({ cwd: ROOT }, options));
|
||||
const cpuCount = os.cpus().length;
|
||||
const totalCount = matchingJSFiles.length;
|
||||
var count = 0;
|
||||
var isError = false;
|
||||
|
||||
cluster.setupMaster({
|
||||
|
@ -48,7 +47,8 @@ async function getJS(source, options, signatures) {
|
|||
const t2 = Date.now();
|
||||
return Promise.resolve({
|
||||
action: 'js',
|
||||
count,
|
||||
count: 0,
|
||||
outFiles: [],
|
||||
totalCount,
|
||||
processingTime: t2 - t1
|
||||
});
|
||||
|
@ -56,6 +56,7 @@ async function getJS(source, options, signatures) {
|
|||
|
||||
// distribute processing among workers
|
||||
const workerCount = Math.min(cpuCount, filesForProcessing.length);
|
||||
const outFiles = [];
|
||||
var workersActive = workerCount;
|
||||
NODE_ENV == 'debug' && console.log(`Will process ${filesForProcessing.length} files using ${workerCount} processes`);
|
||||
return new Promise((resolve, reject) => {
|
||||
|
@ -75,7 +76,7 @@ async function getJS(source, options, signatures) {
|
|||
} else {
|
||||
NODE_ENV == 'debug' && console.log(`process ${this.id} took ${ev.processingTime} ms to process ${ev.sourcefile} into ${ev.outfile}`);
|
||||
NODE_ENV != 'debug' && onProgress(ev.sourcefile, ev.outfile, 'js');
|
||||
count++;
|
||||
outFiles.push(ev.outfile);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -95,7 +96,8 @@ async function getJS(source, options, signatures) {
|
|||
const t2 = Date.now();
|
||||
resolve({
|
||||
action: 'js',
|
||||
count,
|
||||
count: outFiles.length,
|
||||
outFiles,
|
||||
totalCount,
|
||||
processingTime: t2 - t1
|
||||
});
|
||||
|
|
|
@ -11,11 +11,12 @@ const sassRender = universalify.fromCallback(sass.render);
|
|||
|
||||
const ROOT = path.resolve(__dirname, '..');
|
||||
|
||||
async function getSass(source, options, signatures={}) {
|
||||
async function getSass(source, options, signatures = {}) {
|
||||
const t1 = Date.now();
|
||||
const files = await globby(source, Object.assign({ cwd: ROOT }, options));
|
||||
const totalCount = files.length;
|
||||
var count = 0, shouldRebuild = false;
|
||||
const outFiles = [];
|
||||
var shouldRebuild = false;
|
||||
|
||||
for (const f of files) {
|
||||
// if any file changed, rebuild all onSuccess
|
||||
|
@ -54,7 +55,7 @@ async function getSass(source, options, signatures={}) {
|
|||
await fs.outputFile(`${dest}.map`, sass.map);
|
||||
onProgress(f, dest, 'sass');
|
||||
signatures[f] = newFileSignature;
|
||||
count++;
|
||||
outFiles.push(dest);
|
||||
}
|
||||
catch (err) {
|
||||
throw new Error(`Failed on ${f}: ${err}`);
|
||||
|
@ -65,7 +66,8 @@ async function getSass(source, options, signatures={}) {
|
|||
const t2 = Date.now();
|
||||
return {
|
||||
action: 'sass',
|
||||
count,
|
||||
count: outFiles.length,
|
||||
outFiles,
|
||||
totalCount,
|
||||
processingTime: t2 - t1
|
||||
};
|
||||
|
|
|
@ -56,6 +56,7 @@ async function getSymlinks(source, options, signatures) {
|
|||
return {
|
||||
action: 'symlink',
|
||||
count: filesProcessedCount,
|
||||
outFiles: filesToProcess,
|
||||
totalCount: files.length,
|
||||
processingTime: t2 - t1
|
||||
};
|
||||
|
|
|
@ -117,10 +117,13 @@ function comparePaths(actualPath, testedPath) {
|
|||
return path.normalize(actualPath) === path.normalize(testedPath);
|
||||
}
|
||||
|
||||
const envCheckTrue = env => !!(env && (parseInt(env) || env === true || env === "true"));
|
||||
|
||||
module.exports = {
|
||||
cleanUp,
|
||||
comparePaths,
|
||||
compareSignatures,
|
||||
envCheckTrue,
|
||||
formatDirsForMatcher,
|
||||
getFileSignature,
|
||||
getPathRelativeTo,
|
||||
|
|
|
@ -1,8 +1,10 @@
|
|||
const path = require('path');
|
||||
const fs = require('fs-extra');
|
||||
const chokidar = require('chokidar');
|
||||
const multimatch = require('multimatch');
|
||||
const { exec } = require('child_process');
|
||||
const { dirs, jsFiles, scssFiles, ignoreMask, copyDirs, symlinkFiles } = require('./config');
|
||||
const { onSuccess, onError, getSignatures, writeSignatures, cleanUp, formatDirsForMatcher } = require('./utils');
|
||||
const { envCheckTrue, onSuccess, onError, getSignatures, writeSignatures, cleanUp, formatDirsForMatcher } = require('./utils');
|
||||
const getJS = require('./js');
|
||||
const getSass = require('./sass');
|
||||
const getCopy = require('./copy');
|
||||
|
@ -10,6 +12,9 @@ const getSymlinks = require('./symlinks');
|
|||
|
||||
|
||||
const ROOT = path.resolve(__dirname, '..');
|
||||
const addOmniExecPath = path.join(ROOT, '..', 'zotero-standalone-build', 'scripts', 'add_omni_file');
|
||||
let shouldAddOmni = false;
|
||||
|
||||
const source = [
|
||||
'chrome',
|
||||
'components',
|
||||
|
@ -45,32 +50,69 @@ process.on('SIGINT', () => {
|
|||
process.exit();
|
||||
});
|
||||
|
||||
function getWatch() {
|
||||
async function addOmniFiles(relPaths) {
|
||||
const t1 = Date.now();
|
||||
const buildDirPath = path.join(ROOT, 'build');
|
||||
const wrappedPaths = relPaths.map(relPath => `"${path.relative(buildDirPath, relPath)}"`);
|
||||
|
||||
await new Promise((resolve, reject) => {
|
||||
const cmd = `"${addOmniExecPath}" ${wrappedPaths.join(' ')}`;
|
||||
exec(cmd, { cwd: buildDirPath }, (error, output) => {
|
||||
if (error) {
|
||||
reject(error);
|
||||
}
|
||||
else {
|
||||
process.env.NODE_ENV === 'debug' && console.log(`Executed:\n${cmd};\nOutput:\n${output}\n`);
|
||||
resolve(output);
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
const t2 = Date.now();
|
||||
|
||||
return {
|
||||
action: 'add-omni-files',
|
||||
count: relPaths.length,
|
||||
totalCount: relPaths.length,
|
||||
processingTime: t2 - t1
|
||||
};
|
||||
}
|
||||
|
||||
async function getWatch() {
|
||||
try {
|
||||
await fs.access(addOmniExecPath, fs.constants.F_OK);
|
||||
shouldAddOmni = !envCheckTrue(process.env.SKIP_OMNI);
|
||||
}
|
||||
catch (_) {}
|
||||
|
||||
let watcher = chokidar.watch(source, { cwd: ROOT })
|
||||
.on('change', async (path) => {
|
||||
try {
|
||||
var matched = false;
|
||||
var result = false;
|
||||
if (multimatch(path, jsFiles).length && !multimatch(path, ignoreMask).length) {
|
||||
onSuccess(await getJS(path, { ignore: ignoreMask }, signatures));
|
||||
result = await getJS(path, { ignore: ignoreMask }, signatures);
|
||||
onSuccess(await cleanUp(signatures));
|
||||
return;
|
||||
}
|
||||
for (var i = 0; i < scssFiles.length; i++) {
|
||||
if (multimatch(path, scssFiles[i]).length) {
|
||||
onSuccess(await getSass(scssFiles[i], { ignore: ignoreMask }));
|
||||
onSuccess(await cleanUp(signatures));
|
||||
return;
|
||||
if (!result) {
|
||||
for (var i = 0; i < scssFiles.length; i++) {
|
||||
if (multimatch(path, scssFiles[i]).length) {
|
||||
result = await getSass(scssFiles[i], { ignore: ignoreMask }); // eslint-disable-line no-await-in-loop
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
if (multimatch(path, copyDirs.map(d => `${d}/**`)).length) {
|
||||
onSuccess(await getCopy(path, {}, signatures));
|
||||
onSuccess(await cleanUp(signatures));
|
||||
return;
|
||||
if (!result && multimatch(path, copyDirs.map(d => `${d}/**`)).length) {
|
||||
result = await getCopy(path, {}, signatures);
|
||||
}
|
||||
if (multimatch(path, symlinks).length) {
|
||||
onSuccess(await getSymlinks(path, { nodir: true }, signatures));
|
||||
onSuccess(await cleanUp(signatures));
|
||||
return;
|
||||
if (!result && multimatch(path, symlinks).length) {
|
||||
result = await getSymlinks(path, { nodir: true }, signatures);
|
||||
}
|
||||
|
||||
onSuccess(result);
|
||||
onSuccess(await cleanUp(signatures));
|
||||
|
||||
if (shouldAddOmni && result.outFiles?.length) {
|
||||
onSuccess(await addOmniFiles(result.outFiles));
|
||||
}
|
||||
}
|
||||
catch (err) {
|
||||
|
@ -83,7 +125,7 @@ function getWatch() {
|
|||
});
|
||||
|
||||
watcher.add(source);
|
||||
console.log('Watching files for changes...');
|
||||
console.log(`Watching files for changes (omni updates ${shouldAddOmni ? 'enabled' : 'disabled'})...`);
|
||||
}
|
||||
|
||||
module.exports = getWatch;
|
||||
|
@ -93,4 +135,4 @@ if (require.main === module) {
|
|||
signatures = await getSignatures();
|
||||
getWatch();
|
||||
})();
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue