from __future__ import absolute_import, division, print_function import argparse import contextlib import datetime import hashlib import json import os import re import shutil import subprocess import sys import tarfile import tempfile from time import time from multiprocessing import Pool, cpu_count try: import lzma except ImportError: lzma = None def platform_is_win32(): return sys.platform == 'win32' if platform_is_win32(): EXE_SUFFIX = ".exe" else: EXE_SUFFIX = "" def get_cpus(): if hasattr(os, "sched_getaffinity"): return len(os.sched_getaffinity(0)) if hasattr(os, "cpu_count"): cpus = os.cpu_count() if cpus is not None: return cpus try: return cpu_count() except NotImplementedError: return 1 def eprint(*args, **kwargs): kwargs["file"] = sys.stderr print(*args, **kwargs) def get(base, url, path, checksums, verbose=False): with tempfile.NamedTemporaryFile(delete=False) as temp_file: temp_path = temp_file.name try: if url not in checksums: raise RuntimeError(("src/stage0.json doesn't contain a checksum for {}. " "Pre-built artifacts might not be available for this " "target at this time, see https://doc.rust-lang.org/nightly" "/rustc/platform-support.html for more information.") .format(url)) sha256 = checksums[url] if os.path.exists(path): if verify(path, sha256, False): if verbose: eprint("using already-download file", path) return else: if verbose: eprint("ignoring already-download file", path, "due to failed verification") os.unlink(path) download(temp_path, "{}/{}".format(base, url), True, verbose) if not verify(temp_path, sha256, verbose): raise RuntimeError("failed verification") if verbose: eprint("moving {} to {}".format(temp_path, path)) shutil.move(temp_path, path) finally: if os.path.isfile(temp_path): if verbose: eprint("removing", temp_path) os.unlink(temp_path) def download(path, url, probably_big, verbose): for _ in range(4): try: _download(path, url, probably_big, verbose, True) return except RuntimeError: eprint("\nspurious failure, trying again") _download(path, url, probably_big, verbose, False) def _download(path, url, probably_big, verbose, exception): # Try to use curl (potentially available on win32 # https://devblogs.microsoft.com/commandline/tar-and-curl-come-to-windows/) # If an error occurs: # - If we are on win32 fallback to powershell # - Otherwise raise the error if appropriate if probably_big or verbose: eprint("downloading {}".format(url)) try: if (probably_big or verbose) and "GITHUB_ACTIONS" not in os.environ: option = "-#" else: option = "-s" # If curl is not present on Win32, we should not sys.exit # but raise `CalledProcessError` or `OSError` instead require(["curl", "--version"], exception=platform_is_win32()) run(["curl", option, "-L", # Follow redirect. "-y", "30", "-Y", "10", # timeout if speed is < 10 bytes/sec for > 30 seconds "--connect-timeout", "30", # timeout if cannot connect within 30 seconds "-o", path, "--retry", "3", "-SRf", url], verbose=verbose, exception=True, # Will raise RuntimeError on failure ) except (subprocess.CalledProcessError, OSError, RuntimeError): # see http://serverfault.com/questions/301128/how-to-download if platform_is_win32(): run_powershell([ "[Net.ServicePointManager]::SecurityProtocol = [Net.SecurityProtocolType]::Tls12;", "(New-Object System.Net.WebClient).DownloadFile('{}', '{}')".format(url, path)], verbose=verbose, exception=exception) # Check if the RuntimeError raised by run(curl) should be silenced elif verbose or exception: raise def verify(path, expected, verbose): """Check if the sha256 sum of the given path is valid""" if verbose: eprint("verifying", path) with open(path, "rb") as source: found = hashlib.sha256(source.read()).hexdigest() verified = found == expected if not verified: eprint("invalid checksum:\n" " found: {}\n" " expected: {}".format(found, expected)) return verified def unpack(tarball, tarball_suffix, dst, verbose=False, match=None): """Unpack the given tarball file""" eprint("extracting", tarball) fname = os.path.basename(tarball).replace(tarball_suffix, "") with contextlib.closing(tarfile.open(tarball)) as tar: for member in tar.getnames(): if "/" not in member: continue name = member.replace(fname + "/", "", 1) if match is not None and not name.startswith(match): continue name = name[len(match) + 1:] dst_path = os.path.join(dst, name) if verbose: eprint(" extracting", member) tar.extract(member, dst) src_path = os.path.join(dst, member) if os.path.isdir(src_path) and os.path.exists(dst_path): continue shutil.move(src_path, dst_path) shutil.rmtree(os.path.join(dst, fname)) def run(args, verbose=False, exception=False, is_bootstrap=False, **kwargs): """Run a child program in a new process""" if verbose: eprint("running: " + ' '.join(args)) sys.stdout.flush() # Ensure that the .exe is used on Windows just in case a Linux ELF has been # compiled in the same directory. if os.name == 'nt' and not args[0].endswith('.exe'): args[0] += '.exe' # Use Popen here instead of call() as it apparently allows powershell on # Windows to not lock up waiting for input presumably. ret = subprocess.Popen(args, **kwargs) code = ret.wait() if code != 0: err = "failed to run: " + ' '.join(args) if verbose or exception: raise RuntimeError(err) # For most failures, we definitely do want to print this error, or the user will have no # idea what went wrong. But when we've successfully built bootstrap and it failed, it will # have already printed an error above, so there's no need to print the exact command we're # running. if is_bootstrap: sys.exit(1) else: sys.exit(err) def run_powershell(script, *args, **kwargs): """Run a powershell script""" run(["PowerShell.exe", "/nologo", "-Command"] + script, *args, **kwargs) def require(cmd, exit=True, exception=False): '''Run a command, returning its output. On error, If `exception` is `True`, raise the error Otherwise If `exit` is `True`, exit the process Else return None.''' try: return subprocess.check_output(cmd).strip() except (subprocess.CalledProcessError, OSError) as exc: if exception: raise elif exit: eprint("error: unable to run `{}`: {}".format(' '.join(cmd), exc)) eprint("Please make sure it's installed and in the path.") sys.exit(1) return None def format_build_time(duration): """Return a nicer format for build time >>> format_build_time('300') '0:05:00' """ return str(datetime.timedelta(seconds=int(duration))) def default_build_triple(verbose): """Build triple as in LLVM""" # If we're on Windows and have an existing `rustc` toolchain, use `rustc --version --verbose` # to find our host target triple. This fixes an issue with Windows builds being detected # as GNU instead of MSVC. # Otherwise, detect it via `uname` default_encoding = sys.getdefaultencoding() if platform_is_win32(): try: version = subprocess.check_output(["rustc", "--version", "--verbose"], stderr=subprocess.DEVNULL) version = version.decode(default_encoding) host = next(x for x in version.split('\n') if x.startswith("host: ")) triple = host.split("host: ")[1] if verbose: eprint("detected default triple {} from pre-installed rustc".format(triple)) return triple except Exception as e: if verbose: eprint("pre-installed rustc not detected: {}".format(e)) eprint("falling back to auto-detect") required = not platform_is_win32() uname = require(["uname", "-smp"], exit=required) # If we do not have `uname`, assume Windows. if uname is None: return 'x86_64-pc-windows-msvc' kernel, cputype, processor = uname.decode(default_encoding).split(maxsplit=2) # The goal here is to come up with the same triple as LLVM would, # at least for the subset of platforms we're willing to target. kerneltype_mapper = { 'Darwin': 'apple-darwin', 'DragonFly': 'unknown-dragonfly', 'FreeBSD': 'unknown-freebsd', 'Haiku': 'unknown-haiku', 'NetBSD': 'unknown-netbsd', 'OpenBSD': 'unknown-openbsd', 'GNU': 'unknown-hurd', } # Consider the direct transformation first and then the special cases if kernel in kerneltype_mapper: kernel = kerneltype_mapper[kernel] elif kernel == 'Linux': # Apple doesn't support `-o` so this can't be used in the combined # uname invocation above ostype = require(["uname", "-o"], exit=required).decode(default_encoding) if ostype == 'Android': kernel = 'linux-android' else: kernel = 'unknown-linux-gnu' elif kernel == 'SunOS': kernel = 'pc-solaris' # On Solaris, uname -m will return a machine classification instead # of a cpu type, so uname -p is recommended instead. However, the # output from that option is too generic for our purposes (it will # always emit 'i386' on x86/amd64 systems). As such, isainfo -k # must be used instead. cputype = require(['isainfo', '-k']).decode(default_encoding) # sparc cpus have sun as a target vendor if 'sparc' in cputype: kernel = 'sun-solaris' elif kernel.startswith('MINGW'): # msys' `uname` does not print gcc configuration, but prints msys # configuration. so we cannot believe `uname -m`: # msys1 is always i686 and msys2 is always x86_64. # instead, msys defines $MSYSTEM which is MINGW32 on i686 and # MINGW64 on x86_64. kernel = 'pc-windows-gnu' cputype = 'i686' if os.environ.get('MSYSTEM') == 'MINGW64': cputype = 'x86_64' elif kernel.startswith('MSYS'): kernel = 'pc-windows-gnu' elif kernel.startswith('CYGWIN_NT'): cputype = 'i686' if kernel.endswith('WOW64'): cputype = 'x86_64' kernel = 'pc-windows-gnu' elif platform_is_win32(): # Some Windows platforms might have a `uname` command that returns a # non-standard string (e.g. gnuwin32 tools returns `windows32`). In # these cases, fall back to using sys.platform. return 'x86_64-pc-windows-msvc' elif kernel == 'AIX': # `uname -m` returns the machine ID rather than machine hardware on AIX, # so we are unable to use cputype to form triple. AIX 7.2 and # above supports 32-bit and 64-bit mode simultaneously and `uname -p` # returns `powerpc`, however we only supports `powerpc64-ibm-aix` in # rust on AIX. For above reasons, kerneltype_mapper and cputype_mapper # are not used to infer AIX's triple. return 'powerpc64-ibm-aix' else: err = "unknown OS type: {}".format(kernel) sys.exit(err) if cputype in ['powerpc', 'riscv'] and kernel == 'unknown-freebsd': cputype = subprocess.check_output( ['uname', '-p']).strip().decode(default_encoding) cputype_mapper = { 'BePC': 'i686', 'aarch64': 'aarch64', 'aarch64eb': 'aarch64', 'amd64': 'x86_64', 'arm64': 'aarch64', 'i386': 'i686', 'i486': 'i686', 'i686': 'i686', 'i686-AT386': 'i686', 'i786': 'i686', 'loongarch64': 'loongarch64', 'm68k': 'm68k', 'csky': 'csky', 'powerpc': 'powerpc', 'powerpc64': 'powerpc64', 'powerpc64le': 'powerpc64le', 'ppc': 'powerpc', 'ppc64': 'powerpc64', 'ppc64le': 'powerpc64le', 'riscv64': 'riscv64gc', 's390x': 's390x', 'x64': 'x86_64', 'x86': 'i686', 'x86-64': 'x86_64', 'x86_64': 'x86_64' } # Consider the direct transformation first and then the special cases if cputype in cputype_mapper: cputype = cputype_mapper[cputype] elif cputype in {'xscale', 'arm'}: cputype = 'arm' if kernel == 'linux-android': kernel = 'linux-androideabi' elif kernel == 'unknown-freebsd': cputype = processor kernel = 'unknown-freebsd' elif cputype == 'armv6l': cputype = 'arm' if kernel == 'linux-android': kernel = 'linux-androideabi' else: kernel += 'eabihf' elif cputype in {'armv7l', 'armv8l'}: cputype = 'armv7' if kernel == 'linux-android': kernel = 'linux-androideabi' else: kernel += 'eabihf' elif cputype == 'mips': if sys.byteorder == 'big': cputype = 'mips' elif sys.byteorder == 'little': cputype = 'mipsel' else: raise ValueError("unknown byteorder: {}".format(sys.byteorder)) elif cputype == 'mips64': if sys.byteorder == 'big': cputype = 'mips64' elif sys.byteorder == 'little': cputype = 'mips64el' else: raise ValueError('unknown byteorder: {}'.format(sys.byteorder)) # only the n64 ABI is supported, indicate it kernel += 'abi64' elif cputype == 'sparc' or cputype == 'sparcv9' or cputype == 'sparc64': pass else: err = "unknown cpu type: {}".format(cputype) sys.exit(err) return "{}-{}".format(cputype, kernel) @contextlib.contextmanager def output(filepath): tmp = filepath + '.tmp' with open(tmp, 'w') as f: yield f try: if os.path.exists(filepath): os.remove(filepath) # PermissionError/OSError on Win32 if in use except OSError: shutil.copy2(tmp, filepath) os.remove(tmp) return os.rename(tmp, filepath) class Stage0Toolchain: def __init__(self, stage0_payload): self.date = stage0_payload["date"] self.version = stage0_payload["version"] def channel(self): return self.version + "-" + self.date class DownloadInfo: """A helper class that can be pickled into a parallel subprocess""" def __init__( self, base_download_url, download_path, bin_root, tarball_path, tarball_suffix, checksums_sha256, pattern, verbose, ): self.base_download_url = base_download_url self.download_path = download_path self.bin_root = bin_root self.tarball_path = tarball_path self.tarball_suffix = tarball_suffix self.checksums_sha256 = checksums_sha256 self.pattern = pattern self.verbose = verbose def download_component(download_info): if not os.path.exists(download_info.tarball_path): get( download_info.base_download_url, download_info.download_path, download_info.tarball_path, download_info.checksums_sha256, verbose=download_info.verbose, ) def unpack_component(download_info): unpack( download_info.tarball_path, download_info.tarball_suffix, download_info.bin_root, match=download_info.pattern, verbose=download_info.verbose, ) class FakeArgs: """Used for unit tests to avoid updating all call sites""" def __init__(self): self.build = '' self.build_dir = '' self.clean = False self.verbose = False self.json_output = False self.color = 'auto' self.warnings = 'default' class RustBuild(object): """Provide all the methods required to build Rust""" def __init__(self, config_toml="", args=None): if args is None: args = FakeArgs() self.git_version = None self.nix_deps_dir = None self._should_fix_bins_and_dylibs = None self.rust_root = os.path.abspath(os.path.join(__file__, '../../..')) self.config_toml = config_toml self.clean = args.clean self.json_output = args.json_output self.verbose = args.verbose self.color = args.color self.warnings = args.warnings config_verbose_count = self.get_toml('verbose', 'build') if config_verbose_count is not None: self.verbose = max(self.verbose, int(config_verbose_count)) self.use_vendored_sources = self.get_toml('vendor', 'build') == 'true' self.use_locked_deps = self.get_toml('locked-deps', 'build') == 'true' build_dir = args.build_dir or self.get_toml('build-dir', 'build') or 'build' self.build_dir = os.path.abspath(build_dir) with open(os.path.join(self.rust_root, "src", "stage0.json")) as f: data = json.load(f) self.checksums_sha256 = data["checksums_sha256"] self.stage0_compiler = Stage0Toolchain(data["compiler"]) self.download_url = os.getenv("RUSTUP_DIST_SERVER") or data["config"]["dist_server"] self.build = args.build or self.build_triple() def download_toolchain(self): """Fetch the build system for Rust, written in Rust This method will build a cache directory, then it will fetch the tarball which has the stage0 compiler used to then bootstrap the Rust compiler itself. Each downloaded tarball is extracted, after that, the script will move all the content to the right place. """ rustc_channel = self.stage0_compiler.version bin_root = self.bin_root() key = self.stage0_compiler.date if self.rustc().startswith(bin_root) and \ (not os.path.exists(self.rustc()) or self.program_out_of_date(self.rustc_stamp(), key)): if os.path.exists(bin_root): # HACK: On Windows, we can't delete rust-analyzer-proc-macro-server while it's # running. Kill it. if platform_is_win32(): print("Killing rust-analyzer-proc-macro-srv before deleting stage0 toolchain") regex = '{}\\\\(host|{})\\\\stage0\\\\libexec'.format( os.path.basename(self.build_dir), self.build ) script = ( # NOTE: can't use `taskkill` or `Get-Process -Name` because they error if # the server isn't running. 'Get-Process | ' + 'Where-Object {$_.Name -eq "rust-analyzer-proc-macro-srv"} |' + 'Where-Object {{$_.Path -match "{}"}} |'.format(regex) + 'Stop-Process' ) run_powershell([script]) shutil.rmtree(bin_root) key = self.stage0_compiler.date cache_dst = os.path.join(self.build_dir, "cache") rustc_cache = os.path.join(cache_dst, key) if not os.path.exists(rustc_cache): os.makedirs(rustc_cache) tarball_suffix = '.tar.gz' if lzma is None else '.tar.xz' toolchain_suffix = "{}-{}{}".format(rustc_channel, self.build, tarball_suffix) tarballs_to_download = [ ("rust-std-{}".format(toolchain_suffix), "rust-std-{}".format(self.build)), ("rustc-{}".format(toolchain_suffix), "rustc"), ("cargo-{}".format(toolchain_suffix), "cargo"), ] tarballs_download_info = [ DownloadInfo( base_download_url=self.download_url, download_path="dist/{}/{}".format(self.stage0_compiler.date, filename), bin_root=self.bin_root(), tarball_path=os.path.join(rustc_cache, filename), tarball_suffix=tarball_suffix, checksums_sha256=self.checksums_sha256, pattern=pattern, verbose=self.verbose, ) for filename, pattern in tarballs_to_download ] # Download the components serially to show the progress bars properly. for download_info in tarballs_download_info: download_component(download_info) # Unpack the tarballs in parallle. # In Python 2.7, Pool cannot be used as a context manager. pool_size = min(len(tarballs_download_info), get_cpus()) if self.verbose: print('Choosing a pool size of', pool_size, 'for the unpacking of the tarballs') p = Pool(pool_size) try: p.map(unpack_component, tarballs_download_info) finally: p.close() p.join() if self.should_fix_bins_and_dylibs(): self.fix_bin_or_dylib("{}/bin/cargo".format(bin_root)) self.fix_bin_or_dylib("{}/bin/rustc".format(bin_root)) self.fix_bin_or_dylib("{}/bin/rustdoc".format(bin_root)) self.fix_bin_or_dylib("{}/libexec/rust-analyzer-proc-macro-srv".format(bin_root)) lib_dir = "{}/lib".format(bin_root) for lib in os.listdir(lib_dir): if lib.endswith(".so"): self.fix_bin_or_dylib(os.path.join(lib_dir, lib)) with output(self.rustc_stamp()) as rust_stamp: rust_stamp.write(key) def _download_component_helper( self, filename, pattern, tarball_suffix, rustc_cache, ): key = self.stage0_compiler.date tarball = os.path.join(rustc_cache, filename) if not os.path.exists(tarball): get( self.download_url, "dist/{}/{}".format(key, filename), tarball, self.checksums_sha256, verbose=self.verbose, ) unpack(tarball, tarball_suffix, self.bin_root(), match=pattern, verbose=self.verbose) def should_fix_bins_and_dylibs(self): """Whether or not `fix_bin_or_dylib` needs to be run; can only be True on NixOS or if config.toml has `build.patch-binaries-for-nix` set. """ if self._should_fix_bins_and_dylibs is not None: return self._should_fix_bins_and_dylibs def get_answer(): default_encoding = sys.getdefaultencoding() try: ostype = subprocess.check_output( ['uname', '-s']).strip().decode(default_encoding) except subprocess.CalledProcessError: return False except OSError as reason: if getattr(reason, 'winerror', None) is not None: return False raise reason if ostype != "Linux": return False # If the user has explicitly indicated whether binaries should be # patched for Nix, then don't check for NixOS. if self.get_toml("patch-binaries-for-nix", "build") == "true": return True if self.get_toml("patch-binaries-for-nix", "build") == "false": return False # Use `/etc/os-release` instead of `/etc/NIXOS`. # The latter one does not exist on NixOS when using tmpfs as root. try: with open("/etc/os-release", "r") as f: is_nixos = any(ln.strip() in ("ID=nixos", "ID='nixos'", 'ID="nixos"') for ln in f) except FileNotFoundError: is_nixos = False # If not on NixOS, then warn if user seems to be atop Nix shell if not is_nixos: in_nix_shell = os.getenv('IN_NIX_SHELL') if in_nix_shell: eprint("The IN_NIX_SHELL environment variable is `{}`;".format(in_nix_shell), "you may need to set `patch-binaries-for-nix=true` in config.toml") return is_nixos answer = self._should_fix_bins_and_dylibs = get_answer() if answer: eprint("info: You seem to be using Nix.") return answer def fix_bin_or_dylib(self, fname): """Modifies the interpreter section of 'fname' to fix the dynamic linker, or the RPATH section, to fix the dynamic library search path This method is only required on NixOS and uses the PatchELF utility to change the interpreter/RPATH of ELF executables. Please see https://nixos.org/patchelf.html for more information """ assert self._should_fix_bins_and_dylibs is True eprint("attempting to patch", fname) # Only build `.nix-deps` once. nix_deps_dir = self.nix_deps_dir if not nix_deps_dir: # Run `nix-build` to "build" each dependency (which will likely reuse # the existing `/nix/store` copy, or at most download a pre-built copy). # # Importantly, we create a gc-root called `.nix-deps` in the `build/` # directory, but still reference the actual `/nix/store` path in the rpath # as it makes it significantly more robust against changes to the location of # the `.nix-deps` location. # # bintools: Needed for the path of `ld-linux.so` (via `nix-support/dynamic-linker`). # zlib: Needed as a system dependency of `libLLVM-*.so`. # patchelf: Needed for patching ELF binaries (see doc comment above). nix_deps_dir = "{}/{}".format(self.build_dir, ".nix-deps") nix_expr = ''' with (import {}); symlinkJoin { name = "rust-stage0-dependencies"; paths = [ zlib patchelf stdenv.cc.bintools ]; } ''' try: subprocess.check_output([ "nix-build", "-E", nix_expr, "-o", nix_deps_dir, ]) except subprocess.CalledProcessError as reason: eprint("warning: failed to call nix-build:", reason) return self.nix_deps_dir = nix_deps_dir patchelf = "{}/bin/patchelf".format(nix_deps_dir) rpath_entries = [ # Relative default, all binary and dynamic libraries we ship # appear to have this (even when `../lib` is redundant). "$ORIGIN/../lib", os.path.join(os.path.realpath(nix_deps_dir), "lib") ] patchelf_args = ["--set-rpath", ":".join(rpath_entries)] if not fname.endswith(".so"): # Finally, set the correct .interp for binaries with open("{}/nix-support/dynamic-linker".format(nix_deps_dir)) as dynamic_linker: patchelf_args += ["--set-interpreter", dynamic_linker.read().rstrip()] try: subprocess.check_output([patchelf] + patchelf_args + [fname]) except subprocess.CalledProcessError as reason: eprint("warning: failed to call patchelf:", reason) return def rustc_stamp(self): """Return the path for .rustc-stamp at the given stage >>> rb = RustBuild() >>> rb.build = "host" >>> rb.build_dir = "build" >>> expected = os.path.join("build", "host", "stage0", ".rustc-stamp") >>> assert rb.rustc_stamp() == expected, rb.rustc_stamp() """ return os.path.join(self.bin_root(), '.rustc-stamp') def program_out_of_date(self, stamp_path, key): """Check if the given program stamp is out of date""" if not os.path.exists(stamp_path) or self.clean: return True with open(stamp_path, 'r') as stamp: return key != stamp.read() def bin_root(self): """Return the binary root directory for the given stage >>> rb = RustBuild() >>> rb.build = "devel" >>> expected = os.path.abspath(os.path.join("build", "devel", "stage0")) >>> assert rb.bin_root() == expected, rb.bin_root() """ subdir = "stage0" return os.path.join(self.build_dir, self.build, subdir) def get_toml(self, key, section=None): """Returns the value of the given key in config.toml, otherwise returns None >>> rb = RustBuild() >>> rb.config_toml = 'key1 = "value1"\\nkey2 = "value2"' >>> rb.get_toml("key2") 'value2' If the key does not exist, the result is None: >>> rb.get_toml("key3") is None True Optionally also matches the section the key appears in >>> rb.config_toml = '[a]\\nkey = "value1"\\n[b]\\nkey = "value2"' >>> rb.get_toml('key', 'a') 'value1' >>> rb.get_toml('key', 'b') 'value2' >>> rb.get_toml('key', 'c') is None True >>> rb.config_toml = 'key1 = true' >>> rb.get_toml("key1") 'true' """ return RustBuild.get_toml_static(self.config_toml, key, section) @staticmethod def get_toml_static(config_toml, key, section=None): cur_section = None for line in config_toml.splitlines(): section_match = re.match(r'^\s*\[(.*)\]\s*$', line) if section_match is not None: cur_section = section_match.group(1) match = re.match(r'^{}\s*=(.*)$'.format(key), line) if match is not None: value = match.group(1) if section is None or section == cur_section: return RustBuild.get_string(value) or value.strip() return None def cargo(self): """Return config path for cargo""" return self.program_config('cargo') def rustc(self): """Return config path for rustc""" return self.program_config('rustc') def program_config(self, program): """Return config path for the given program at the given stage >>> rb = RustBuild() >>> rb.config_toml = 'rustc = "rustc"\\n' >>> rb.program_config('rustc') 'rustc' >>> rb.config_toml = '' >>> cargo_path = rb.program_config('cargo') >>> cargo_path.rstrip(".exe") == os.path.join(rb.bin_root(), ... "bin", "cargo") True """ config = self.get_toml(program) if config: return os.path.expanduser(config) return os.path.join(self.bin_root(), "bin", "{}{}".format(program, EXE_SUFFIX)) @staticmethod def get_string(line): """Return the value between double quotes >>> RustBuild.get_string(' "devel" ') 'devel' >>> RustBuild.get_string(" 'devel' ") 'devel' >>> RustBuild.get_string('devel') is None True >>> RustBuild.get_string(' "devel ') '' """ start = line.find('"') if start != -1: end = start + 1 + line[start + 1:].find('"') return line[start + 1:end] start = line.find('\'') if start != -1: end = start + 1 + line[start + 1:].find('\'') return line[start + 1:end] return None def bootstrap_binary(self): """Return the path of the bootstrap binary >>> rb = RustBuild() >>> rb.build_dir = "build" >>> rb.bootstrap_binary() == os.path.join("build", "bootstrap", ... "debug", "bootstrap") True """ return os.path.join(self.build_dir, "bootstrap", "debug", "bootstrap") def build_bootstrap(self): """Build bootstrap""" env = os.environ.copy() if "GITHUB_ACTIONS" in env: print("::group::Building bootstrap") else: eprint("Building bootstrap") args = self.build_bootstrap_cmd(env) # Run this from the source directory so cargo finds .cargo/config run(args, env=env, verbose=self.verbose, cwd=self.rust_root) if "GITHUB_ACTIONS" in env: print("::endgroup::") def build_bootstrap_cmd(self, env): """For tests.""" build_dir = os.path.join(self.build_dir, "bootstrap") if self.clean and os.path.exists(build_dir): shutil.rmtree(build_dir) # `CARGO_BUILD_TARGET` breaks bootstrap build. # See also: . if "CARGO_BUILD_TARGET" in env: del env["CARGO_BUILD_TARGET"] env["CARGO_TARGET_DIR"] = build_dir env["RUSTC"] = self.rustc() env["LD_LIBRARY_PATH"] = os.path.join(self.bin_root(), "lib") + \ (os.pathsep + env["LD_LIBRARY_PATH"]) \ if "LD_LIBRARY_PATH" in env else "" env["DYLD_LIBRARY_PATH"] = os.path.join(self.bin_root(), "lib") + \ (os.pathsep + env["DYLD_LIBRARY_PATH"]) \ if "DYLD_LIBRARY_PATH" in env else "" env["LIBRARY_PATH"] = os.path.join(self.bin_root(), "lib") + \ (os.pathsep + env["LIBRARY_PATH"]) \ if "LIBRARY_PATH" in env else "" env["LIBPATH"] = os.path.join(self.bin_root(), "lib") + \ (os.pathsep + env["LIBPATH"]) \ if "LIBPATH" in env else "" # Export Stage0 snapshot compiler related env variables build_section = "target.{}".format(self.build) host_triple_sanitized = self.build.replace("-", "_") var_data = { "CC": "cc", "CXX": "cxx", "LD": "linker", "AR": "ar", "RANLIB": "ranlib" } for var_name, toml_key in var_data.items(): toml_val = self.get_toml(toml_key, build_section) if toml_val is not None: env["{}_{}".format(var_name, host_triple_sanitized)] = toml_val # preserve existing RUSTFLAGS env.setdefault("RUSTFLAGS", "") target_features = [] if self.get_toml("crt-static", build_section) == "true": target_features += ["+crt-static"] elif self.get_toml("crt-static", build_section) == "false": target_features += ["-crt-static"] if target_features: env["RUSTFLAGS"] += " -C target-feature=" + (",".join(target_features)) target_linker = self.get_toml("linker", build_section) if target_linker is not None: env["RUSTFLAGS"] += " -C linker=" + target_linker env["RUSTFLAGS"] += " -Wrust_2018_idioms -Wunused_lifetimes" if self.warnings == "default": deny_warnings = self.get_toml("deny-warnings", "rust") != "false" else: deny_warnings = self.warnings == "deny" if deny_warnings: env["RUSTFLAGS"] += " -Dwarnings" env["PATH"] = os.path.join(self.bin_root(), "bin") + \ os.pathsep + env["PATH"] if not os.path.isfile(self.cargo()): raise Exception("no cargo executable found at `{}`".format( self.cargo())) args = [self.cargo(), "build", "--manifest-path", os.path.join(self.rust_root, "src/bootstrap/Cargo.toml")] args.extend("--verbose" for _ in range(self.verbose)) if self.use_locked_deps: args.append("--locked") if self.use_vendored_sources: args.append("--frozen") if self.get_toml("metrics", "build"): args.append("--features") args.append("build-metrics") if self.json_output: args.append("--message-format=json") if self.color == "always": args.append("--color=always") elif self.color == "never": args.append("--color=never") try: args += env["CARGOFLAGS"].split() except KeyError: pass return args def build_triple(self): """Build triple as in LLVM Note that `default_build_triple` is moderately expensive, so use `self.build` where possible. """ config = self.get_toml('build') return config or default_build_triple(self.verbose) def check_vendored_status(self): """Check that vendoring is configured properly""" # keep this consistent with the equivalent check in rustbuild: # https://github.com/rust-lang/rust/blob/a8a33cf27166d3eabaffc58ed3799e054af3b0c6/src/bootstrap/lib.rs#L399-L405 if 'SUDO_USER' in os.environ and not self.use_vendored_sources: if os.getuid() == 0: self.use_vendored_sources = True eprint('info: looks like you\'re trying to run this command as root') eprint(' and so in order to preserve your $HOME this will now') eprint(' use vendored sources by default.') cargo_dir = os.path.join(self.rust_root, '.cargo') if self.use_vendored_sources: vendor_dir = os.path.join(self.rust_root, 'vendor') if not os.path.exists(vendor_dir): sync_dirs = "--sync ./src/tools/cargo/Cargo.toml " \ "--sync ./src/tools/rust-analyzer/Cargo.toml " \ "--sync ./compiler/rustc_codegen_cranelift/Cargo.toml " \ "--sync ./src/bootstrap/Cargo.toml " eprint('error: vendoring required, but vendor directory does not exist.') eprint(' Run `cargo vendor {}` to initialize the ' 'vendor directory.'.format(sync_dirs)) eprint('Alternatively, use the pre-vendored `rustc-src` dist component.') raise Exception("{} not found".format(vendor_dir)) if not os.path.exists(cargo_dir): eprint('error: vendoring required, but .cargo/config does not exist.') raise Exception("{} not found".format(cargo_dir)) else: if os.path.exists(cargo_dir): shutil.rmtree(cargo_dir) def parse_args(args): """Parse the command line arguments that the python script needs.""" parser = argparse.ArgumentParser(add_help=False) parser.add_argument('-h', '--help', action='store_true') parser.add_argument('--config') parser.add_argument('--build-dir') parser.add_argument('--build') parser.add_argument('--color', choices=['always', 'never', 'auto']) parser.add_argument('--clean', action='store_true') parser.add_argument('--json-output', action='store_true') parser.add_argument('--warnings', choices=['deny', 'warn', 'default'], default='default') parser.add_argument('-v', '--verbose', action='count', default=0) return parser.parse_known_args(args)[0] def bootstrap(args): """Configure, fetch, build and run the initial bootstrap""" rust_root = os.path.abspath(os.path.join(__file__, '../../..')) # Read from `--config`, then `RUST_BOOTSTRAP_CONFIG`, then `./config.toml`, # then `config.toml` in the root directory. toml_path = args.config or os.getenv('RUST_BOOTSTRAP_CONFIG') using_default_path = toml_path is None if using_default_path: toml_path = 'config.toml' if not os.path.exists(toml_path): toml_path = os.path.join(rust_root, toml_path) # Give a hard error if `--config` or `RUST_BOOTSTRAP_CONFIG` are set to a missing path, # but not if `config.toml` hasn't been created. if not using_default_path or os.path.exists(toml_path): with open(toml_path) as config: config_toml = config.read() else: config_toml = '' profile = RustBuild.get_toml_static(config_toml, 'profile') if profile is not None: # Allows creating alias for profile names, allowing # profiles to be renamed while maintaining back compatibility # Keep in sync with `profile_aliases` in config.rs profile_aliases = { "user": "dist" } include_file = 'config.{}.toml'.format(profile_aliases.get(profile) or profile) include_dir = os.path.join(rust_root, 'src', 'bootstrap', 'defaults') include_path = os.path.join(include_dir, include_file) # HACK: This works because `self.get_toml()` returns the first match it finds for a # specific key, so appending our defaults at the end allows the user to override them with open(include_path) as included_toml: config_toml += os.linesep + included_toml.read() # Configure initial bootstrap build = RustBuild(config_toml, args) build.check_vendored_status() if not os.path.exists(build.build_dir): os.makedirs(build.build_dir) # Fetch/build the bootstrap build.download_toolchain() sys.stdout.flush() build.build_bootstrap() sys.stdout.flush() # Run the bootstrap args = [build.bootstrap_binary()] args.extend(sys.argv[1:]) env = os.environ.copy() env["BOOTSTRAP_PARENT_ID"] = str(os.getpid()) env["BOOTSTRAP_PYTHON"] = sys.executable run(args, env=env, verbose=build.verbose, is_bootstrap=True) def main(): """Entry point for the bootstrap process""" start_time = time() # x.py help ... if len(sys.argv) > 1 and sys.argv[1] == 'help': sys.argv[1] = '-h' args = parse_args(sys.argv) help_triggered = args.help or len(sys.argv) == 1 # If the user is asking for help, let them know that the whole download-and-build # process has to happen before anything is printed out. if help_triggered: eprint( "info: Downloading and building bootstrap before processing --help command.\n" " See src/bootstrap/README.md for help with common commands.") exit_code = 0 success_word = "successfully" try: bootstrap(args) except (SystemExit, KeyboardInterrupt) as error: if hasattr(error, 'code') and isinstance(error.code, int): exit_code = error.code else: exit_code = 1 eprint(error) success_word = "unsuccessfully" if not help_triggered: eprint("Build completed", success_word, "in", format_build_time(time() - start_time)) sys.exit(exit_code) if __name__ == '__main__': main()