mirror of
https://github.com/SukkaW/Surge.git
synced 2025-12-12 01:00:34 +08:00
Chore: download dist from GitLab if GitHub dist repo is down
This commit is contained in:
parent
affaa13890
commit
554afd0fd9
3
.github/workflows/main.yml
vendored
3
.github/workflows/main.yml
vendored
@ -74,8 +74,8 @@ jobs:
|
||||
run: |
|
||||
mkdir ./deploy-git
|
||||
cd ./deploy-git
|
||||
git init
|
||||
git config --global init.defaultBranch master
|
||||
git init
|
||||
git config --global push.default matching
|
||||
git config --global user.email "${GITLAB_EMAIL}"
|
||||
git config --global user.name "${GITLAB_USER}"
|
||||
@ -95,6 +95,7 @@ jobs:
|
||||
GITLAB_TOKEN: ${{ secrets.GITLAB_TOKEN }}
|
||||
- name: Upload Dist to GitHub
|
||||
uses: peaceiris/actions-gh-pages@v4
|
||||
continue-on-error: true
|
||||
with:
|
||||
personal_token: ${{ secrets.GIT_TOKEN }}
|
||||
user_name: ${{ secrets.GIT_USER }}
|
||||
|
||||
@ -13,6 +13,9 @@ import { Readable } from 'node:stream';
|
||||
const IS_READING_BUILD_OUTPUT = 1 << 2;
|
||||
const ALL_FILES_EXISTS = 1 << 3;
|
||||
|
||||
const GITHUB_CODELOAD_URL = 'https://codeload.github.com/sukkalab/ruleset.skk.moe/tar.gz/master';
|
||||
const GITLAB_CODELOAD_URL = 'https://gitlab.com/SukkaW/ruleset.skk.moe/-/archive/master/ruleset.skk.moe-master.tar.gz';
|
||||
|
||||
export const downloadPreviousBuild = task(require.main === module, __filename)(async (span) => {
|
||||
const buildOutputList: string[] = [];
|
||||
|
||||
@ -49,54 +52,71 @@ export const downloadPreviousBuild = task(require.main === module, __filename)(a
|
||||
|
||||
const filesList = buildOutputList.map(f => path.join('ruleset.skk.moe-master', f));
|
||||
|
||||
return span
|
||||
.traceChild('download & extract previoud build')
|
||||
.traceAsyncFn(async () => {
|
||||
const resp = await fetchWithRetry('https://codeload.github.com/sukkalab/ruleset.skk.moe/tar.gz/master', {
|
||||
...defaultRequestInit,
|
||||
retry: {
|
||||
retryOnNon2xx: false
|
||||
}
|
||||
});
|
||||
|
||||
if (resp.status !== 200) {
|
||||
console.warn('Download previous build failed! Status:', resp.status);
|
||||
if (resp.status === 404) {
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
if (!resp.body) {
|
||||
throw new Error('Download previous build failed! No body found');
|
||||
}
|
||||
|
||||
const gunzip = zlib.createGunzip();
|
||||
const extract = tarStream.extract();
|
||||
|
||||
pipeline(
|
||||
Readable.fromWeb(resp.body),
|
||||
gunzip,
|
||||
extract
|
||||
);
|
||||
|
||||
const pathPrefix = `ruleset.skk.moe-master${path.sep}`;
|
||||
|
||||
for await (const entry of extract) {
|
||||
if (entry.header.type !== 'file') {
|
||||
entry.resume(); // Drain the entry
|
||||
continue;
|
||||
}
|
||||
// filter entry
|
||||
if (!filesList.some(f => entry.header.name.startsWith(f))) {
|
||||
entry.resume(); // Drain the entry
|
||||
continue;
|
||||
}
|
||||
|
||||
const relativeEntryPath = entry.header.name.replace(pathPrefix, '');
|
||||
const targetPath = path.join(__dirname, '..', relativeEntryPath);
|
||||
|
||||
await mkdir(path.dirname(targetPath), { recursive: true });
|
||||
await pipeline(entry, createWriteStream(targetPath));
|
||||
const tarGzUrl = await span.traceChildAsync('get tar.gz url', async () => {
|
||||
const resp = await fetchWithRetry(GITHUB_CODELOAD_URL, {
|
||||
...defaultRequestInit,
|
||||
method: 'HEAD',
|
||||
retry: {
|
||||
retryOnNon2xx: false
|
||||
}
|
||||
});
|
||||
if (resp.status !== 200) {
|
||||
console.warn('Download previous build from GitHub failed! Status:', resp.status);
|
||||
console.warn('Switch to GitLab');
|
||||
return GITLAB_CODELOAD_URL;
|
||||
}
|
||||
return GITHUB_CODELOAD_URL;
|
||||
});
|
||||
|
||||
return span.traceChildAsync('download & extract previoud build', async () => {
|
||||
const resp = await fetchWithRetry(tarGzUrl, {
|
||||
headers: {
|
||||
'User-Agent': 'curl/8.9.1',
|
||||
Accept: 'application/octet-stream'
|
||||
},
|
||||
retry: {
|
||||
retryOnNon2xx: false
|
||||
}
|
||||
});
|
||||
|
||||
if (resp.status !== 200) {
|
||||
console.warn('Download previous build failed! Status:', resp.status);
|
||||
if (resp.status === 404) {
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
if (!resp.body) {
|
||||
throw new Error('Download previous build failed! No body found');
|
||||
}
|
||||
|
||||
const gunzip = zlib.createGunzip();
|
||||
const extract = tarStream.extract();
|
||||
|
||||
pipeline(
|
||||
Readable.fromWeb(resp.body),
|
||||
gunzip,
|
||||
extract
|
||||
);
|
||||
|
||||
const pathPrefix = 'ruleset.skk.moe-master/';
|
||||
|
||||
for await (const entry of extract) {
|
||||
if (entry.header.type !== 'file') {
|
||||
entry.resume(); // Drain the entry
|
||||
continue;
|
||||
}
|
||||
// filter entry
|
||||
if (!filesList.some(f => entry.header.name.startsWith(f))) {
|
||||
entry.resume(); // Drain the entry
|
||||
continue;
|
||||
}
|
||||
|
||||
const relativeEntryPath = entry.header.name.replace(pathPrefix, '');
|
||||
const targetPath = path.join(__dirname, '..', relativeEntryPath);
|
||||
|
||||
await mkdir(path.dirname(targetPath), { recursive: true });
|
||||
await pipeline(entry, createWriteStream(targetPath));
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
@ -124,7 +124,7 @@ function createFetchRetry($fetch: typeof fetch): FetchWithRetry {
|
||||
|
||||
export const defaultRequestInit: RequestInit = {
|
||||
headers: {
|
||||
'User-Agent': 'curl/8.9.0 (https://github.com/SukkaW/Surge)'
|
||||
'User-Agent': 'curl/8.9.1 (https://github.com/SukkaW/Surge)'
|
||||
}
|
||||
};
|
||||
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user