123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641 |
- #!/usr/bin/env python
- import argparse
- import collections
- import contextlib
- import datetime
- import io
- import json
- import logging
- import os
- from pathlib import Path
- import platform
- import re
- import shutil
- import subprocess
- import sys
- import tarfile
- import tempfile
- import textwrap
- import typing
- import zipfile
- logger = logging.getLogger(__name__)
- VcArchDevel = collections.namedtuple("VcArchDevel", ("dll", "imp", "test"))
- GIT_HASH_FILENAME = ".git-hash"
- def itertools_batched(iterator: typing.Iterable, count: int):
- iterator = iter(iterator)
- while True:
- items = []
- for _ in range(count):
- obj = next(iterator, None)
- if obj is None:
- yield tuple(items)
- return
- items.append(obj)
- yield tuple(items)
- class Executer:
- def __init__(self, root: Path, dry: bool=False):
- self.root = root
- self.dry = dry
- def run(self, cmd, stdout=False, dry_out=None, force=False):
- sys.stdout.flush()
- logger.info("Executing args=%r", cmd)
- if self.dry and not force:
- if stdout:
- return subprocess.run(["echo", "-E", dry_out or ""], stdout=subprocess.PIPE if stdout else None, text=True, check=True, cwd=self.root)
- else:
- return subprocess.run(cmd, stdout=subprocess.PIPE if stdout else None, text=True, check=True, cwd=self.root)
- class SectionPrinter:
- @contextlib.contextmanager
- def group(self, title: str):
- print(f"{title}:")
- yield
- class GitHubSectionPrinter(SectionPrinter):
- def __init__(self):
- super().__init__()
- self.in_group = False
- @contextlib.contextmanager
- def group(self, title: str):
- print(f"::group::{title}")
- assert not self.in_group, "Can enter a group only once"
- self.in_group = True
- yield
- self.in_group = False
- print("::endgroup::")
- class VisualStudio:
- def __init__(self, executer: Executer, year: typing.Optional[str]=None):
- self.executer = executer
- self.vsdevcmd = self.find_vsdevcmd(year)
- self.msbuild = self.find_msbuild()
- @property
- def dry(self):
- return self.executer.dry
- VS_YEAR_TO_VERSION = {
- "2022": 17,
- "2019": 16,
- "2017": 15,
- "2015": 14,
- "2013": 12,
- }
- def find_vsdevcmd(self, year: typing.Optional[str]=None) -> typing.Optional[Path]:
- vswhere_spec = ["-latest"]
- if year is not None:
- try:
- version = cls.VS_YEAR_TO_VERSION[year]
- except KeyError:
- logger.error("Invalid Visual Studio year")
- return None
- vswhere_spec.extend(["-version", f"[{version},{version+1})"])
- vswhere_cmd = ["vswhere"] + vswhere_spec + ["-property", "installationPath"]
- vs_install_path = Path(self.executer.run(vswhere_cmd, stdout=True, dry_out="/tmp").stdout.strip())
- logger.info("VS install_path = %s", vs_install_path)
- assert vs_install_path.is_dir(), "VS installation path does not exist"
- vsdevcmd_path = vs_install_path / "Common7/Tools/vsdevcmd.bat"
- logger.info("vsdevcmd path = %s", vsdevcmd_path)
- if self.dry:
- vsdevcmd_path.parent.mkdir(parents=True, exist_ok=True)
- vsdevcmd_path.touch(exist_ok=True)
- assert vsdevcmd_path.is_file(), "vsdevcmd.bat batch file does not exist"
- return vsdevcmd_path
- def find_msbuild(self) -> typing.Optional[Path]:
- vswhere_cmd = ["vswhere", "-latest", "-requires", "Microsoft.Component.MSBuild", "-find", "MSBuild\**\Bin\MSBuild.exe"]
- msbuild_path = Path(self.executer.run(vswhere_cmd, stdout=True, dry_out="/tmp/MSBuild.exe").stdout.strip())
- logger.info("MSBuild path = %s", msbuild_path)
- if self.dry:
- msbuild_path.parent.mkdir(parents=True, exist_ok=True)
- msbuild_path.touch(exist_ok=True)
- assert msbuild_path.is_file(), "MSBuild.exe does not exist"
- return msbuild_path
- def build(self, arch: str, platform: str, configuration: str, projects: list[Path]):
- assert projects, "Need at least one project to build"
- vsdev_cmd_str = f"\"{self.vsdevcmd}\" -arch={arch}"
- msbuild_cmd_str = " && ".join([f"\"{self.msbuild}\" \"{project}\" /m /p:BuildInParallel=true /p:Platform={platform} /p:Configuration={configuration}" for project in projects])
- bat_contents = f"{vsdev_cmd_str} && {msbuild_cmd_str}\n"
- bat_path = Path(tempfile.gettempdir()) / "cmd.bat"
- with bat_path.open("w") as f:
- f.write(bat_contents)
- logger.info("Running cmd.exe script (%s): %s", bat_path, bat_contents)
- cmd = ["cmd.exe", "/D", "/E:ON", "/V:OFF", "/S", "/C", f"CALL {str(bat_path)}"]
- self.executer.run(cmd)
- class Releaser:
- def __init__(self, project: str, commit: str, root: Path, dist_path: Path, section_printer: SectionPrinter, executer: Executer, cmake_generator: str):
- self.project = project
- self.version = self.extract_sdl_version(root=root, project=project)
- self.root = root
- self.commit = commit
- self.dist_path = dist_path
- self.section_printer = section_printer
- self.executer = executer
- self.cmake_generator = cmake_generator
- self.artifacts = {}
- @property
- def dry(self):
- return self.executer.dry
- def prepare(self):
- logger.debug("Creating dist folder")
- self.dist_path.mkdir(parents=True, exist_ok=True)
- GitLsTreeResult = collections.namedtuple("GitLsTreeResult", ("path", "mode", "object_type", "object_name"))
- def _git_ls_tree(self, commit) -> dict[str, GitLsTreeResult]:
- logger.debug("Getting source listing from git")
- dry_out = textwrap.dedent("""\
- "CMakeLists.txt": {"object_name": "9e5e4bcf094bfbde94f19c3f314808031ec8f141", "mode": "100644", "type": "blob"},
- """)
- last_key = "zzzzzz"
- dict_tree_items = "{" + self.executer.run(["git", "ls-tree", "-r", """--format="%(path)": {"object_name": "%(objectname)", "mode": "%(objectmode)", "type": "%(objecttype)"},""", commit], stdout=True, dry_out=dry_out).stdout + f'"{last_key}": null' + "}"
- with open("/tmp/a.txt", "w") as f:
- f.write(dict_tree_items)
- f.write("\n")
- dict_tree_items = json.loads(dict_tree_items)
- del dict_tree_items[last_key]
- tree_items = {path: self.GitLsTreeResult(path=path, mode=int(v["mode"], 8), object_type=v["type"], object_name=v["object_name"]) for path, v in dict_tree_items.items()}
- assert all(item.object_type == "blob" for item in tree_items.values())
- return tree_items
- def _git_cat_file(self, tree_items: dict[str, GitLsTreeResult]) -> dict[str, bytes]:
- logger.debug("Getting source binary data from git")
- if self.dry:
- return {
- "CMakeLists.txt": b"cmake_minimum_required(VERSION 3.20)\nproject(SDL)\n",
- }
- git_cat = subprocess.Popen(["git", "cat-file", "--batch"], stdin=subprocess.PIPE, stdout=subprocess.PIPE, text=False, bufsize=50 * 1024 * 1024)
- data_tree = {}
- batch_size = 60
- for batch in itertools_batched(tree_items.items(), batch_size):
- for object_path, tree_item in batch:
- logger.debug("Requesting data of file '%s' (object=%s)...", object_path, tree_item.object_name)
- git_cat.stdin.write(f"{tree_item.object_name}\n".encode())
- git_cat.stdin.flush()
- for object_path, tree_item in batch:
- header = git_cat.stdout.readline().decode()
- object_name, object_type, obj_size = header.strip().split(maxsplit=3)
- assert tree_item.object_name == object_name
- assert tree_item.object_type == object_type
- obj_size = int(obj_size)
- data_tree[object_path] = git_cat.stdout.read(obj_size)
- logger.debug("File data received '%s'", object_path)
- assert git_cat.stdout.readline() == b"\n"
- assert len(data_tree) == len(tree_items)
- logger.debug("No more file!")
- git_cat.stdin.close()
- git_cat.wait()
- assert git_cat.returncode == 0
- logger.debug("All data received!")
- return data_tree
- def _get_file_times(self, tree_items: dict[str, GitLsTreeResult]) -> dict[str, datetime.datetime]:
- dry_out = textwrap.dedent("""\
- time=2024-03-14T15:40:25-07:00
- M\tCMakeLists.txt
- """)
- git_log_out = self.executer.run(["git", "log", "--name-status", '--pretty=time=%cI'], stdout=True, dry_out=dry_out).stdout.splitlines(keepends=False)
- current_time = None
- tree_paths = {item.path for item in tree_items.values()}
- path_times = {}
- for line in git_log_out:
- if not line:
- continue
- if line.startswith("time="):
- current_time = datetime.datetime.fromisoformat(line.removeprefix("time="))
- continue
- mod_type, paths = line.split(maxsplit=1)
- assert current_time is not None
- for path in paths.split():
- if path in tree_paths and path not in path_times:
- path_times[path] = current_time
- assert set(path_times.keys()) == tree_paths
- return path_times
- @staticmethod
- def _path_filter(path: str):
- if path.startswith(".git"):
- return False
- return True
- TreeItem = collections.namedtuple("TreeItem", ("path", "mode", "data", "time"))
- def _get_git_contents(self) -> dict[str, (TreeItem, bytes, datetime.datetime)]:
- commit_file_tree = self._git_ls_tree(self.commit)
- git_datas = self._git_cat_file(commit_file_tree)
- git_times = self._get_file_times(commit_file_tree)
- git_contents = {path: self.TreeItem(path=path, data=git_datas[path], mode=item.mode, time=git_times[path]) for path, item in commit_file_tree.items() if self._path_filter(path)}
- return git_contents
- def create_source_archives(self):
- archive_base = f"{self.project}-{self.version}"
- git_contents = self._get_git_contents()
- git_files = list(git_contents.values())
- assert len(git_contents) == len(git_files)
- latest_mod_time = max(item.time for item in git_files)
- git_files.append(self.TreeItem(path="VERSION.txt", data=f"{self.version}\n".encode(), mode=0o100644, time=latest_mod_time))
- git_files.append(self.TreeItem(path=GIT_HASH_FILENAME, data=f"{self.commit}\n".encode(), mode=0o100644, time=latest_mod_time))
- git_files.sort(key=lambda v: v.time)
- zip_path = self.dist_path / f"{archive_base}.zip"
- logger.info("Creating .zip source archive (%s)...", zip_path)
- if self.dry:
- zip_path.touch()
- else:
- with zipfile.ZipFile(zip_path, "w", compression=zipfile.ZIP_DEFLATED) as zip_object:
- for git_file in git_files:
- file_data_time = (git_file.time.year, git_file.time.month, git_file.time.day, git_file.time.hour, git_file.time.minute, git_file.time.second)
- zip_info = zipfile.ZipInfo(filename=f"{archive_base}/{git_file.path}", date_time=file_data_time)
- zip_info.external_attr = git_file.mode << 16
- zip_info.compress_type = zipfile.ZIP_DEFLATED
- zip_object.writestr(zip_info, data=git_file.data)
- self.artifacts["src-zip"] = zip_path
- tar_types = (
- (".tar.gz", "gz"),
- (".tar.xz", "xz"),
- )
- for ext, comp in tar_types:
- tar_path = self.dist_path / f"{archive_base}{ext}"
- logger.info("Creating %s source archive (%s)...", ext, tar_path)
- if self.dry:
- tar_path.touch()
- else:
- with tarfile.open(tar_path, f"w:{comp}") as tar_object:
- for git_file in git_files:
- tar_info = tarfile.TarInfo(f"{archive_base}/{git_file.path}")
- tar_info.mode = git_file.mode
- tar_info.size = len(git_file.data)
- tar_info.mtime = git_file.time.timestamp()
- tar_object.addfile(tar_info, fileobj=io.BytesIO(git_file.data))
- if tar_path.suffix == ".gz":
- # Zero the embedded timestamp in the gzip'ed tarball
- with open(tar_path, "r+b") as f:
- f.seek(4, 0)
- f.write(b"\x00\x00\x00\x00")
- self.artifacts[f"src-tar-{comp}"] = tar_path
- def create_xcframework(self, configuration: str="Release"):
- dmg_in = self.root / f"Xcode/SDL/build/SDL3.dmg"
- dmg_in.unlink(missing_ok=True)
- self.executer.run(["xcodebuild", "-project", self.root / "Xcode/SDL/SDL.xcodeproj", "-target", "SDL3.dmg", "-configuration", configuration])
- if self.dry:
- dmg_in.parent.mkdir(parents=True, exist_ok=True)
- dmg_in.touch()
- assert dmg_in.is_file(), "SDL3.dmg was not created by xcodebuild"
- dmg_out = self.dist_path / f"{self.project}-{self.version}.dmg"
- shutil.copy(dmg_in, dmg_out)
- self.artifacts["dmg"] = dmg_out
- @property
- def git_hash_data(self):
- return f"{self.commit}\n".encode()
- def _tar_add_git_hash(self, tar_object: tarfile.TarFile, root: typing.Optional[str]=None, time: typing.Optional[datetime.datetime]=None):
- if not time:
- time = datetime.datetime(year=2024, month=4, day=1)
- path = GIT_HASH_FILENAME
- if root:
- path = f"{root}/{path}"
- tar_info = tarfile.TarInfo(path)
- tar_info.mode = 0o100644
- tar_info.size = len(self.git_hash_data)
- tar_info.mtime = time.timestamp()
- tar_object.addfile(tar_info, fileobj=io.BytesIO(self.git_hash_data))
- def _zip_add_git_hash(self, zip_file: zipfile.ZipFile, root: typing.Optional[str]=None, time: typing.Optional[datetime.datetime]=None):
- if not time:
- time = datetime.datetime(year=2024, month=4, day=1)
- path = GIT_HASH_FILENAME
- if root:
- path = f"{root}/{path}"
- file_data_time = (time.year, time.month, time.day, time.hour, time.minute, time.second)
- zip_info = zipfile.ZipInfo(filename=path, date_time=file_data_time)
- zip_info.external_attr = 0o100644 << 16
- zip_info.compress_type = zipfile.ZIP_DEFLATED
- zip_file.writestr(zip_info, data=self.git_hash_data)
- def create_mingw_archives(self):
- build_type = "Release"
- mingw_archs = ("i686", "x86_64")
- build_parent_dir = self.root / "build-mingw"
- zip_path = self.dist_path / f"{self.project}-devel-{self.version}-mingw.zip"
- tar_exts = ("gz", "xz")
- tar_paths = { ext: self.dist_path / f"{self.project}-devel-{self.version}-mingw.tar.{ext}" for ext in tar_exts}
- arch_install_paths = {}
- arch_files = {}
- for arch in mingw_archs:
- build_path = build_parent_dir / f"build-{arch}"
- install_path = build_parent_dir / f"install-{arch}"
- arch_install_paths[arch] = install_path
- shutil.rmtree(install_path, ignore_errors=True)
- build_path.mkdir(parents=True, exist_ok=True)
- with self.section_printer.group(f"Configuring MinGW {arch}"):
- self.executer.run([
- "cmake", "-S", str(self.root), "-B", str(build_path),
- "--fresh",
- "-DSDL_SHARED=ON",
- "-DSDL_STATIC=ON",
- "-DSDL_DISABLE_INSTALL_DOCS=ON",
- "-DSDL_TEST_LIBRARY=ON",
- "-DSDL_TESTS=OFF",
- "-DCMAKE_INSTALL_BINDIR=bin",
- "-DCMAKE_INSTALL_DATAROOTDIR=share",
- "-DCMAKE_INSTALL_INCLUDEDIR=include",
- "-DCMAKE_INSTALL_LIBDIR=lib",
- f"-DCMAKE_BUILD_TYPE={build_type}",
- f"-DCMAKE_TOOLCHAIN_FILE={self.root}/build-scripts/cmake-toolchain-mingw64-{arch}.cmake",
- f"-G{self.cmake_generator}",
- f"-DCMAKE_INSTALL_PREFIX={install_path}",
- ])
- with self.section_printer.group(f"Build MinGW {arch}"):
- self.executer.run(["cmake", "--build", str(build_path), "--verbose", "--config", build_type])
- with self.section_printer.group(f"Install MinGW {arch}"):
- self.executer.run(["cmake", "--install", str(build_path), "--strip", "--config", build_type])
- arch_files[arch] = list(Path(r) / f for r, _, files in os.walk(install_path) for f in files)
- extra_files = [
- ("mingw/pkg-support/INSTALL.txt", ""),
- ("mingw/pkg-support/Makefile", ""),
- ("mingw/pkg-support/cmake/sdl3-config.cmake", "cmake/"),
- ("mingw/pkg-support/cmake/sdl3-config-version.cmake", "cmake/"),
- ("BUGS.txt", ""),
- ("CREDITS.md", ""),
- ("README-SDL.txt", ""),
- ("WhatsNew.txt", ""),
- ("LICENSE.txt", ""),
- ("README.md", ""),
- ]
- test_files = list(Path(r) / f for r, _, files in os.walk(self.root / "test") for f in files)
- # FIXME: split SDL3.dll debug information into debug library
- # objcopy --only-keep-debug SDL3.dll SDL3.debug.dll
- # objcopy --add-gnu-debuglink=SDL3.debug.dll SDL3.dll
- # objcopy --strip-debug SDL3.dll
- for comp in tar_exts:
- logger.info("Creating %s...", tar_paths[comp])
- with tarfile.open(tar_paths[comp], f"w:{comp}") as tar_object:
- arc_root = f"{self.project}-{self.version}"
- for file_path, arcdirname in extra_files:
- assert not arcdirname or arcdirname[-1] == "/"
- arcname = f"{arc_root}/{arcdirname}{Path(file_path).name}"
- tar_object.add(self.root / file_path, arcname=arcname)
- for arch in mingw_archs:
- install_path = arch_install_paths[arch]
- arcname_parent = f"{arc_root}/{arch}-w64-mingw32"
- for file in arch_files[arch]:
- arcname = os.path.join(arcname_parent, file.relative_to(install_path))
- tar_object.add(file, arcname=arcname)
- for test_file in test_files:
- arcname = f"{arc_root}/test/{test_file.relative_to(self.root/'test')}"
- tar_object.add(test_file, arcname=arcname)
- self._tar_add_git_hash(tar_object=tar_object, root=arc_root)
- self.artifacts[f"mingw-devel-tar-{comp}"] = tar_paths[comp]
- def build_vs(self, arch: str, platform: str, vs: VisualStudio, configuration: str="Release"):
- dll_path = self.root / f"VisualC/SDL/{platform}/{configuration}/{self.project}.dll"
- imp_path = self.root / f"VisualC/SDL/{platform}/{configuration}/{self.project}.lib"
- test_path = self.root / f"VisualC/SDL_test/{platform}/{configuration}/{self.project}_test.lib"
- dll_path.unlink(missing_ok=True)
- imp_path.unlink(missing_ok=True)
- test_path.unlink(missing_ok=True)
- projects = [
- self.root / "VisualC/SDL/SDL.vcxproj",
- self.root / "VisualC/SDL_test/SDL_test.vcxproj",
- ]
- vs.build(arch=arch, platform=platform, configuration=configuration, projects=projects)
- if self.dry:
- dll_path.parent.mkdir(parents=True, exist_ok=True)
- dll_path.touch()
- imp_path.touch()
- test_path.parent.mkdir(parents=True, exist_ok=True)
- test_path.touch()
- assert dll_path.is_file(), "SDL3.dll has not been created"
- assert imp_path.is_file(), "SDL3.lib has not been created"
- assert test_path.is_file(), "SDL3_test.lib has not been created"
- zip_path = self.dist_path / f"{self.project}-{self.version}-win32-{arch}.zip"
- zip_path.unlink(missing_ok=True)
- logger.info("Creating %s", zip_path)
- with zipfile.ZipFile(zip_path, mode="w", compression=zipfile.ZIP_DEFLATED) as zf:
- logger.debug("Adding %s", dll_path.name)
- zf.write(dll_path, arcname=dll_path.name)
- logger.debug("Adding %s", "README-SDL.txt")
- zf.write(self.root / "README-SDL.txt", arcname="README-SDL.txt")
- self._zip_add_git_hash(zip_file=zf)
- self.artifacts[f"VC-{arch}"] = zip_path
- return VcArchDevel(dll=dll_path, imp=imp_path, test=test_path)
- def build_vs_devel(self, arch_vc: dict[str, VcArchDevel]):
- zip_path = self.dist_path / f"{self.project}-devel-{self.version}-VC.zip"
- archive_prefix = f"{self.project}-{self.version}"
- def zip_file(zf: zipfile.ZipFile, path: Path, arcrelpath: str):
- arcname = f"{archive_prefix}/{arcrelpath}"
- logger.debug("Adding %s to %s", path, arcname)
- zf.write(path, arcname=arcname)
- def zip_directory(zf: zipfile.ZipFile, directory: Path, arcrelpath: str):
- for f in directory.iterdir():
- if f.is_file():
- arcname = f"{archive_prefix}/{arcrelpath}/{f.name}"
- logger.debug("Adding %s to %s", f, arcname)
- zf.write(f, arcname=arcname)
- with zipfile.ZipFile(zip_path, mode="w", compression=zipfile.ZIP_DEFLATED) as zf:
- for arch, binaries in arch_vc.items():
- zip_file(zf, path=binaries.dll, arcrelpath=f"lib/{arch}/{binaries.dll.name}")
- zip_file(zf, path=binaries.imp, arcrelpath=f"lib/{arch}/{binaries.imp.name}")
- zip_file(zf, path=binaries.test, arcrelpath=f"lib/{arch}/{binaries.test.name}")
- zip_directory(zf, directory=self.root / "include/SDL3", arcrelpath="include/SDL3")
- zip_directory(zf, directory=self.root / "docs", arcrelpath="docs")
- zip_directory(zf, directory=self.root / "VisualC/pkg-support/cmake", arcrelpath="cmake")
- for txt in ("BUGS.txt", "README-SDL.txt", "WhatsNew.txt"):
- zip_file(zf, path=self.root / txt, arcrelpath=txt)
- zip_file(zf, path=self.root / "LICENSE.txt", arcrelpath="COPYING.txt")
- zip_file(zf, path=self.root / "README.md", arcrelpath="README.txt")
- self._zip_add_git_hash(zip_file=zf, root=archive_prefix)
- self.artifacts["VC-devel"] = zip_path
- @classmethod
- def extract_sdl_version(cls, root: Path, project: str):
- with open(root / f"include/{project}/SDL_version.h", "r") as f:
- text = f.read()
- major = next(re.finditer(r"^#define SDL_MAJOR_VERSION\s+([0-9]+)$", text, flags=re.M)).group(1)
- minor = next(re.finditer(r"^#define SDL_MINOR_VERSION\s+([0-9]+)$", text, flags=re.M)).group(1)
- patch = next(re.finditer(r"^#define SDL_PATCHLEVEL\s+([0-9]+)$", text, flags=re.M)).group(1)
- return f"{major}.{minor}.{patch}"
- def main(argv=None):
- parser = argparse.ArgumentParser(allow_abbrev=False, description="Create SDL release artifacts")
- parser.add_argument("--root", metavar="DIR", type=Path, default=Path(__file__).resolve().parents[1], help="Root of SDL")
- parser.add_argument("--out", "-o", metavar="DIR", dest="dist_path", type=Path, default="dist", help="Output directory")
- parser.add_argument("--github", action="store_true", help="Script is running on a GitHub runner")
- parser.add_argument("--commit", default="HEAD", help="Git commit/tag of which a release should be created")
- parser.add_argument("--project", required=True, help="Name of the project")
- parser.add_argument("--create", choices=["source", "mingw", "win32", "xcframework"], required=True,action="append", dest="actions", help="SDL version")
- parser.set_defaults(loglevel=logging.INFO)
- parser.add_argument('--vs-year', dest="vs_year", help="Visual Studio year")
- parser.add_argument('--cmake-generator', dest="cmake_generator", default="Ninja", help="CMake Generator")
- parser.add_argument('--debug', action='store_const', const=logging.DEBUG, dest="loglevel", help="Print script debug information")
- parser.add_argument('--dry-run', action='store_true', dest="dry", help="Don't execute anything")
- parser.add_argument('--force', action='store_true', dest="force", help="Ignore a non-clean git tree")
- args = parser.parse_args(argv)
- logging.basicConfig(level=args.loglevel, format='[%(levelname)s] %(message)s')
- args.actions = set(args.actions)
- args.dist_path = args.dist_path.resolve()
- args.root = args.root.resolve()
- args.dist_path = args.dist_path.resolve()
- if args.dry:
- args.dist_path = args.dist_path / "dry"
- if args.github:
- section_printer = GitHubSectionPrinter()
- else:
- section_printer = SectionPrinter()
- executer = Executer(root=args.root, dry=args.dry)
- root_git_hash_path = args.root / GIT_HASH_FILENAME
- root_is_maybe_archive = root_git_hash_path.is_file()
- if root_is_maybe_archive:
- logger.warning("%s detected: Building from archive", GIT_HASH_FILENAME)
- archive_commit = root_git_hash_path.read_text().strip()
- if args.commit != archive_commit:
- logger.warn("Commit argument is %s, but archive commit is %s. Using %s.", args.commit, archive_commit, archive_commit)
- args.commit = archive_commit
- else:
- args.commit = executer.run(["git", "rev-parse", args.commit], stdout=True, dry_out="e5812a9fd2cda317b503325a702ba3c1c37861d9").stdout.strip()
- logger.info("Using commit %s", args.commit)
- releaser = Releaser(
- project=args.project,
- commit=args.commit,
- root=args.root,
- dist_path=args.dist_path,
- executer=executer,
- section_printer=section_printer,
- cmake_generator=args.cmake_generator,
- )
- if root_is_maybe_archive:
- logger.warn("Building from archive. Skipping clean git tree check.")
- else:
- porcelain_status = executer.run(["git", "status", "--ignored", "--porcelain"], stdout=True, dry_out="\n").stdout.strip()
- if porcelain_status:
- print(porcelain_status)
- logger.warning("The tree is dirty! Do not publish any generated artifacts!")
- if not args.force:
- raise Exception("The git repo contains modified and/or non-committed files. Run with --force to ignore.")
- with section_printer.group("Arguments"):
- print(f"project = {args.project}")
- print(f"version = {releaser.version}")
- print(f"commit = {args.commit}")
- print(f"out = {args.dist_path}")
- print(f"actions = {args.actions}")
- print(f"dry = {args.dry}")
- print(f"force = {args.force}")
- print(f"cmake_generator = {args.cmake_generator}")
- releaser.prepare()
- if "source" in args.actions:
- if root_is_maybe_archive:
- raise Exception("Cannot build source archive from source archive")
- with section_printer.group("Create source archives"):
- releaser.create_source_archives()
- if "xcframework" in args.actions:
- if platform.system() != "Darwin" and not args.dry:
- parser.error("xcframework artifact(s) can only be built on Darwin")
- releaser.create_xcframework()
- if "win32" in args.actions:
- if platform.system() != "Windows" and not args.dry:
- parser.error("win32 artifact(s) can only be built on Windows")
- with section_printer.group("Find Visual Studio"):
- vs = VisualStudio(executer=executer)
- with section_printer.group("Build x86 VS binary"):
- x86 = releaser.build_vs(arch="x86", platform="Win32", vs=vs)
- with section_printer.group("Build x64 VS binary"):
- x64 = releaser.build_vs(arch="x64", platform="x64", vs=vs)
- with section_printer.group("Create SDL VC development zip"):
- arch_vc = {
- "x86": x86,
- "x64": x64,
- }
- releaser.build_vs_devel(arch_vc)
- if "mingw" in args.actions:
- releaser.create_mingw_archives()
- with section_printer.group("Summary"):
- print(f"artifacts = {releaser.artifacts}")
- if args.github:
- if args.dry:
- os.environ["GITHUB_OUTPUT"] = "/tmp/github_output.txt"
- with open(os.environ["GITHUB_OUTPUT"], "a") as f:
- f.write(f"project={releaser.project}\n")
- f.write(f"version={releaser.version}\n")
- for k, v in releaser.artifacts.items():
- f.write(f"{k}={v.name}\n")
- return 0
- if __name__ == "__main__":
- raise SystemExit(main())
|