Multi architecture descriptors
Some checks failed
Continuous integration / test (push) Has been cancelled

This commit is contained in:
Vladimir Borisov 2025-12-05 15:44:01 +02:00
parent fd84baff33
commit 0a08fed2cb
No known key found for this signature in database
GPG key ID: F9A584BE4FCB6603
2 changed files with 94 additions and 72 deletions

View file

@ -1,4 +0,0 @@
[target.x86_64-pc-windows-msvc]
rustflags = ["-C", "link-args=/LIBPATH:.\\mpv-x64"]
[target.aarch64-pc-windows-msvc]
rustflags = ["-C", "link-args=/LIBPATH:.\\mpv-arm64"]

View file

@ -21,6 +21,10 @@ const DOWNLOAD_ENDPOINT = "https://dl.strem.io/stremio-shell-ng/";
const OS_EXT = { const OS_EXT = {
".exe": "windows", ".exe": "windows",
}; };
const ARCH = [
"x64",
"arm64",
];
const VERSION_REGEX = /^v(\d+\.\d+\.\d+).*$/; const VERSION_REGEX = /^v(\d+\.\d+\.\d+).*$/;
const supportedArguments = Object.freeze({ const supportedArguments = Object.freeze({
@ -99,7 +103,7 @@ const parseArguments = () => {
return acc; return acc;
}, {}); }, {});
try { try {
for (let i = 2; i < process.argv.length; i++) { for (let i = 2;i < process.argv.length;i++) {
const arg = process.argv[i]; const arg = process.argv[i];
if (arg.startsWith("--")) { if (arg.startsWith("--")) {
// Stop processing arguments after -- // Stop processing arguments after --
@ -146,6 +150,7 @@ const s3Cp = (src, dest) => s3Cmd(`cp --acl public-read ${src} ${dest}`);
// Downloads a file from S3 and returns a hash of it // Downloads a file from S3 and returns a hash of it
const s3Hash = (path) => { const s3Hash = (path) => {
log("Downloading ", path)
return new Promise((resolve, reject) => { return new Promise((resolve, reject) => {
const hash = createHash("sha256"); const hash = createHash("sha256");
https https
@ -167,12 +172,17 @@ const parseS3Listing = (tag) => (line) => {
) || {} ) || {}
).groups; ).groups;
if (!path) return; if (!path) return;
const os = OS_EXT[posix.extname(path.name)]; const ext = posix.extname(path.name);
const os = OS_EXT[ext];
const fn = posix.basename(path.name, ext)
if (!os) return; if (!os) return;
const arch = ARCH.find(s => fn.endsWith('_' + s))
return { return {
name: path.name, name: path.name,
url: `${DOWNLOAD_ENDPOINT + tag}/${path.name}`, url: `${DOWNLOAD_ENDPOINT + tag}/${path.name}`,
os, os,
arch,
tag,
date: new Date(path.date), date: new Date(path.date),
}; };
}; };
@ -185,6 +195,12 @@ const getFilesForTag = async (tag) =>
}) })
).filter((file) => file); ).filter((file) => file);
const groupBy = (prop, groups, item) => {
Array.isArray(groups[item[prop]]) ? groups[item[prop]].push(item) : groups[item[prop]] = [item]
return groups
}
const calculateFileChecksums = async (files) => const calculateFileChecksums = async (files) =>
Promise.all( Promise.all(
files.map(async (file) => { files.map(async (file) => {
@ -197,6 +213,50 @@ const calculateFileChecksums = async (files) =>
}) })
); );
const uploadDescriptor = async (descriptor, release, force, dry_run) => {
const suffix = descriptor.arch ? '_' + descriptor.arch : ''
const desc_name = descriptor.tag + suffix + ".json";
const s3_rc_desc_path = S3_VERSIONS_RC_PATH + desc_name;
const s3_dest_path = release
? S3_VERSIONS_PATH + desc_name
: s3_rc_desc_path;
const rc_uploaded = await s3Ls(s3_dest_path).then((listing) => !!listing)
if (!force && rc_uploaded) {
log(`${descriptor.arch || 'noarch'} ${args.release ? "" : "RC "}escriptor for tag ${descriptor.tag} already exists. Skipping upload.`)
return
}
if (
release &&
!force &&
rc_uploaded
) {
log(
"Descriptor for tag",
tag,
"already exists in the RC folder. Moving it to the releases folder"
);
if (!dry_run) await s3Cp(s3_rc_desc_path, s3_dest_path);
log("Done");
return;
}
const descriptor_path = `${tmpdir()}/${desc_name}`;
const descriptor_text = JSON.stringify(descriptor, null, 2) + "\n";
log("Writting descriptor to", descriptor_path);
if (dry_run) log(descriptor_text)
else fs.writeFileSync(descriptor_path, descriptor_text);
log(`Uploading ${descriptor.arch || 'noarch'} ${release ? "" : "RC "}descriptor to S3`);
try {
if (!dry_run) await s3Cp(descriptor_path, s3_dest_path);
} finally {
// Clean up the temporary file even if the upload fails
log("Cleaning up", descriptor_path);
if (!dry_run) fs.unlinkSync(descriptor_path);
}
log("Done uploading", descriptor_path, "to", s3_dest_path);
};
// Generates the descriptor for a given tag // Generates the descriptor for a given tag
// If no tag is provided, it will get the latest tag // If no tag is provided, it will get the latest tag
// An example descriptor: // An example descriptor:
@ -208,92 +268,58 @@ const calculateFileChecksums = async (files) =>
// "files": [ // "files": [
// { // {
// "name": "StremioSetup.exe", // "name": "StremioSetup.exe",
// "url": "https://s3-eu-west-1.amazonaws.com/stremio-artifacts/stremio-shell-ng/v0.1.0-new-setup/StremioSetup.exe", // "url": "https://s3-eu-west-1.amazonaws.com/stremio-artifacts/stremio-shell-ng/v0.1.0-new-setup/Stremio.exe",
// "checksum": "0ff94905df4d94233d14f48ed68e31664a478a29204be4c7867c2389929c6ac3", // "checksum": "0ff94905df4d94233d14f48ed68e31664a478a29204be4c7867c2389929c6ac3",
// "os": "windows" // "os": "macos"
// } // }
// ] // ]
// } // }
const generateDescriptor = async (args) => { const generateDescriptors = async (args) => {
let tag = args.tag; let tag = args.tag;
if (!tag) { if (!tag) {
log("Obtaining the latest tag"); log("Obtaining the latest tag");
tag = await s3Ls(S3_BUCKET_PATH).then((listing) => { tag = await s3Ls(S3_BUCKET_PATH).then((listing) => {
// get the first line, remove the DIR prefix, and get the basename // get the last line, remove the DIR prefix, and get the basename
// which is the tag // which is the tag
const first_path = listing.replace(/^\s+\w+\s+/gm, '').split('\n').find(line => line.match(VERSION_REGEX)); const ver_path = listing.replace(/^\s+\w+\s+/gm, '').split('\n').reverse().find(line => line.match(VERSION_REGEX));
return posix.basename(first_path); return posix.basename(ver_path);
}); });
} }
const desc_name = tag + ".json";
const s3_rc_desc_path = S3_VERSIONS_RC_PATH + desc_name;
const s3_dest_path = args.release
? S3_VERSIONS_PATH + desc_name
: s3_rc_desc_path;
if (!args.force && await s3Ls(s3_dest_path).then((listing) => !!listing)) {
throw new Error(
`${args.release ? "" : "RC "}Descriptor for tag ${tag} already exists`
);
}
if (
args.release &&
!args.force &&
(await s3Ls(s3_rc_desc_path).then((listing) => !!listing))
) {
log(
"Descriptor for tag",
tag,
"already exists in the RC folder. Moving it to the releases folder"
);
if (!args.dry_run) await s3Cp(s3_rc_desc_path, s3_dest_path);
log("Done");
return;
}
log("Getting files for tag", tag); log("Getting files for tag", tag);
if (!tag) throw new Error("No tag found"); if (!tag) throw new Error("No tag found");
const version = (tag.match(VERSION_REGEX) || [])[1]; const version = (tag.match(VERSION_REGEX) || [])[1];
if (!version) throw new Error("No valid version found"); if (!version) throw new Error("No valid version found");
const file_listing = await getFilesForTag(tag); const arch_listing = (await getFilesForTag(tag)).reduce(groupBy.bind(null, 'arch'), {});
// We need at least one file to extract the release date (await Promise.all(Object.keys(arch_listing).map(async arch => {
if (!file_listing.length) throw new Error("No files found"); const file_listing = arch_listing[arch]
if (args.wait_all && file_listing.length < Object.keys(OS_EXT).length) { // We need at least one file to extract the release date
log( if (!file_listing.length) throw new Error("No files found");
`Not all files are uploaded yet. Rerun this script after ${Object.keys(OS_EXT).length - file_listing.length
} more are uploaded`
);
return;
}
const files = await calculateFileChecksums(file_listing);
const descriptor = {
version,
tag,
released: file_listing[0].date.toISOString(),
files,
};
const descriptor_text = JSON.stringify(descriptor, null, 2) + "\n";
if (args.dry_run) {
process.stdout.write(descriptor_text);
return;
}
const descriptor_path = `${tmpdir()}/${desc_name}`; const files = await calculateFileChecksums(file_listing);
log("Writting descriptor to", descriptor_path); const descriptor = {
fs.writeFileSync(descriptor_path, descriptor_text); version,
tag,
log(`Uploading ${args.release ? "" : "RC "}descriptor to S3`); released: file_listing[0].date.toISOString(),
try { arch: file_listing[0].arch,
await s3Cp(descriptor_path, s3_dest_path); files: files.map(f => ({
} finally { url: f.url,
// Clean up the temporary file even if the upload fails checksum: f.checksum,
log("Cleaning up"); os: f.os,
fs.unlinkSync(descriptor_path); date: f.date.toISOString(),
} })),
log("Done"); };
uploadDescriptor(descriptor, args.release, args.force, args.dry_run);
if (descriptor.arch == 'x64') {
descriptor.arch = '';
uploadDescriptor(descriptor, args.release, args.force, args.dry_run);
}
return descriptor
})));
}; };
generateDescriptor(args).catch((err) => { generateDescriptors(args).catch((err) => {
console.error(err.message); console.error(err.message);
process.exit(1); process.exit(1);
}); });