mirror of
https://github.com/SukkaW/Surge.git
synced 2025-12-12 01:00:34 +08:00
110 lines
2.9 KiB
JavaScript
110 lines
2.9 KiB
JavaScript
const { fetch } = require('undici');
|
|
const tar = require('tar');
|
|
const fs = require('fs');
|
|
const fse = require('fs-extra');
|
|
const { join, resolve } = require('path');
|
|
const { tmpdir } = require('os');
|
|
const { Readable } = require('stream');
|
|
const { pipeline } = require('stream/promises');
|
|
const { readFileByLine } = require('./lib/fetch-remote-text-by-line');
|
|
const { isCI } = require('ci-info');
|
|
const { task, traceAsync } = require('./lib/trace-runner');
|
|
|
|
const fileExists = (path) => {
|
|
return fs.promises.access(path, fs.constants.F_OK)
|
|
.then(() => true)
|
|
.catch(() => false);
|
|
};
|
|
|
|
const downloadPreviousBuild = task(__filename, async () => {
|
|
const filesList = ['Clash', 'List'];
|
|
|
|
let allFileExists = true;
|
|
|
|
for await (const line of readFileByLine(resolve(__dirname, '../.gitignore'))) {
|
|
if (
|
|
(
|
|
// line.startsWith('List/')
|
|
line.startsWith('Modules/')
|
|
) && !line.endsWith('/')
|
|
) {
|
|
filesList.push(line);
|
|
|
|
if (!isCI) {
|
|
allFileExists = fs.existsSync(join(__dirname, '..', line));
|
|
if (!allFileExists) {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (isCI) {
|
|
allFileExists = false;
|
|
}
|
|
|
|
if (allFileExists) {
|
|
console.log('All files exists, skip download.');
|
|
return;
|
|
}
|
|
|
|
const extractedPath = join(tmpdir(), `sukka-surge-last-build-extracted-${Date.now()}`);
|
|
|
|
await traceAsync(
|
|
'Download and extract previous build',
|
|
() => Promise.all([
|
|
fetch('https://codeload.github.com/sukkaw/surge/tar.gz/gh-pages'),
|
|
fse.ensureDir(extractedPath)
|
|
]).then(([resp]) => pipeline(
|
|
Readable.fromWeb(resp.body),
|
|
tar.x({
|
|
cwd: extractedPath,
|
|
filter(p) {
|
|
const dir = p.split('/')[1];
|
|
return dir === 'List' || dir === 'Modules' || dir === 'Clash';
|
|
}
|
|
})
|
|
))
|
|
);
|
|
|
|
console.log('Files list:', filesList);
|
|
|
|
await Promise.all(filesList.map(async p => {
|
|
const src = join(extractedPath, 'Surge-gh-pages', p);
|
|
if (await fileExists(src)) {
|
|
return fse.copy(
|
|
src,
|
|
join(__dirname, '..', p),
|
|
{ overwrite: true }
|
|
);
|
|
}
|
|
}));
|
|
|
|
return fs.promises.unlink(extractedPath).catch(() => { });
|
|
});
|
|
|
|
const downloadPublicSuffixList = task(__filename, async () => {
|
|
const publicSuffixDir = resolve(__dirname, '../node_modules/.cache');
|
|
const publicSuffixPath = join(publicSuffixDir, 'public_suffix_list_dat.txt');
|
|
|
|
const [resp] = await Promise.all([
|
|
fetch('https://publicsuffix.org/list/public_suffix_list.dat'),
|
|
fse.ensureDir(publicSuffixDir)
|
|
]);
|
|
|
|
return pipeline(
|
|
Readable.fromWeb(resp.body),
|
|
fs.createWriteStream(publicSuffixPath)
|
|
);
|
|
}, 'download-publicsuffixlist');
|
|
|
|
module.exports.downloadPreviousBuild = downloadPreviousBuild;
|
|
module.exports.downloadPublicSuffixList = downloadPublicSuffixList;
|
|
|
|
if (require.main === module) {
|
|
Promise.all([
|
|
downloadPreviousBuild(),
|
|
downloadPublicSuffixList()
|
|
]);
|
|
}
|