mirror of
https://github.com/SukkaW/Surge.git
synced 2025-12-12 01:00:34 +08:00
CI: push to new repo
This commit is contained in:
parent
f867f3e3d1
commit
f550f0bd68
5
.github/workflows/main.yml
vendored
5
.github/workflows/main.yml
vendored
@ -31,9 +31,10 @@ jobs:
|
|||||||
- name: Deploy
|
- name: Deploy
|
||||||
uses: peaceiris/actions-gh-pages@v3
|
uses: peaceiris/actions-gh-pages@v3
|
||||||
with:
|
with:
|
||||||
github_token: ${{ secrets.GIT_TOKEN }}
|
personal_token: ${{ secrets.GIT_TOKEN }}
|
||||||
user_name: ${{ secrets.GIT_USER }}
|
user_name: ${{ secrets.GIT_USER }}
|
||||||
user_email: ${{ secrets.GIT_EMAIL }}
|
user_email: ${{ secrets.GIT_EMAIL }}
|
||||||
publish_branch: gh-pages
|
external_repository: SukkaLab/ruleset.skk.moe
|
||||||
|
publish_branch: master
|
||||||
publish_dir: public
|
publish_dir: public
|
||||||
cname: ruleset.skk.moe
|
cname: ruleset.skk.moe
|
||||||
|
|||||||
@ -1,8 +1,9 @@
|
|||||||
const { fetch } = require('undici');
|
const { fetch } = require('undici');
|
||||||
const tar = require('tar');
|
const tar = require('tar');
|
||||||
const fs = require('fs');
|
const fs = require('fs');
|
||||||
|
const fsp = fs.promises;
|
||||||
const fse = require('fs-extra');
|
const fse = require('fs-extra');
|
||||||
const { join, resolve } = require('path');
|
const path = require('path');
|
||||||
const { tmpdir } = require('os');
|
const { tmpdir } = require('os');
|
||||||
const { Readable } = require('stream');
|
const { Readable } = require('stream');
|
||||||
const { pipeline } = require('stream/promises');
|
const { pipeline } = require('stream/promises');
|
||||||
@ -21,7 +22,7 @@ const downloadPreviousBuild = task(__filename, async () => {
|
|||||||
|
|
||||||
let allFileExists = true;
|
let allFileExists = true;
|
||||||
|
|
||||||
for await (const line of readFileByLine(resolve(__dirname, '../.gitignore'))) {
|
for await (const line of readFileByLine(path.resolve(__dirname, '../.gitignore'))) {
|
||||||
if (
|
if (
|
||||||
(
|
(
|
||||||
// line.startsWith('List/')
|
// line.startsWith('List/')
|
||||||
@ -31,7 +32,7 @@ const downloadPreviousBuild = task(__filename, async () => {
|
|||||||
filesList.push(line);
|
filesList.push(line);
|
||||||
|
|
||||||
if (!isCI) {
|
if (!isCI) {
|
||||||
allFileExists = fs.existsSync(join(__dirname, '..', line));
|
allFileExists = fs.existsSync(path.join(__dirname, '..', line));
|
||||||
if (!allFileExists) {
|
if (!allFileExists) {
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
@ -48,20 +49,19 @@ const downloadPreviousBuild = task(__filename, async () => {
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const extractedPath = join(tmpdir(), `sukka-surge-last-build-extracted-${Date.now()}`);
|
const extractedPath = path.join(tmpdir(), `sukka-surge-last-build-extracted-${Date.now()}`);
|
||||||
|
|
||||||
await traceAsync(
|
await traceAsync(
|
||||||
'Download and extract previous build',
|
'Download and extract previous build',
|
||||||
() => Promise.all([
|
() => Promise.all([
|
||||||
fetch('https://codeload.github.com/sukkaw/surge/tar.gz/gh-pages'),
|
fetch('https://codeload.github.com/sukkalab/ruleset.skk.moe/tar.gz/master'),
|
||||||
fse.ensureDir(extractedPath)
|
fsp.mkdir(extractedPath, { recursive: true })
|
||||||
]).then(([resp]) => pipeline(
|
]).then(([resp]) => pipeline(
|
||||||
Readable.fromWeb(resp.body),
|
Readable.fromWeb(resp.body),
|
||||||
tar.x({
|
tar.x({
|
||||||
cwd: extractedPath,
|
cwd: extractedPath,
|
||||||
filter(p) {
|
filter(p) {
|
||||||
const dir = p.split('/')[1];
|
return p.includes('/List/') || p.includes('/Modules/') || p.includes('/Clash/');
|
||||||
return dir === 'List' || dir === 'Modules' || dir === 'Clash';
|
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
))
|
))
|
||||||
@ -70,26 +70,26 @@ const downloadPreviousBuild = task(__filename, async () => {
|
|||||||
console.log('Files list:', filesList);
|
console.log('Files list:', filesList);
|
||||||
|
|
||||||
await Promise.all(filesList.map(async p => {
|
await Promise.all(filesList.map(async p => {
|
||||||
const src = join(extractedPath, 'Surge-gh-pages', p);
|
const src = path.join(extractedPath, 'ruleset.skk.moe-master', p);
|
||||||
if (await fileExists(src)) {
|
if (await fileExists(src)) {
|
||||||
return fse.copy(
|
return fse.copy(
|
||||||
src,
|
src,
|
||||||
join(__dirname, '..', p),
|
path.join(__dirname, '..', p),
|
||||||
{ overwrite: true }
|
{ overwrite: true }
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
}));
|
}));
|
||||||
|
|
||||||
return fs.promises.unlink(extractedPath).catch(() => { });
|
// return fs.promises.unlink(extractedPath).catch(() => { });
|
||||||
});
|
});
|
||||||
|
|
||||||
const downloadPublicSuffixList = task(__filename, async () => {
|
const downloadPublicSuffixList = task(__filename, async () => {
|
||||||
const publicSuffixDir = resolve(__dirname, '../node_modules/.cache');
|
const publicSuffixDir = path.resolve(__dirname, '../node_modules/.cache');
|
||||||
const publicSuffixPath = join(publicSuffixDir, 'public_suffix_list_dat.txt');
|
const publicSuffixPath = path.join(publicSuffixDir, 'public_suffix_list_dat.txt');
|
||||||
|
|
||||||
const [resp] = await Promise.all([
|
const [resp] = await Promise.all([
|
||||||
fetch('https://publicsuffix.org/list/public_suffix_list.dat'),
|
fetch('https://publicsuffix.org/list/public_suffix_list.dat'),
|
||||||
fse.ensureDir(publicSuffixDir)
|
fsp.mkdir(publicSuffixDir, { recursive: true })
|
||||||
]);
|
]);
|
||||||
|
|
||||||
return pipeline(
|
return pipeline(
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user