//! Implementation of compiling various phases of the compiler and standard //! library. //! //! This module contains some of the real meat in the rustbuild build system //! which is where Cargo is used to compile the standard library, libtest, and //! the compiler. This module is also responsible for assembling the sysroot as it //! goes along from the output of the previous stage. use std::borrow::Cow; use std::collections::HashSet; use std::env; use std::ffi::OsStr; use std::fs; use std::io::prelude::*; use std::io::BufReader; use std::path::{Path, PathBuf}; use std::process::{Command, Stdio}; use std::str; use serde_derive::Deserialize; use crate::builder::crate_description; use crate::builder::Cargo; use crate::builder::{Builder, Kind, PathSet, RunConfig, ShouldRun, Step, TaskPath}; use crate::cache::{Interned, INTERNER}; use crate::config::{DebuginfoLevel, LlvmLibunwind, RustcLto, TargetSelection}; use crate::dist; use crate::llvm; use crate::tool::SourceType; use crate::util::get_clang_cl_resource_dir; use crate::util::{exe, is_debug_info, is_dylib, output, symlink_dir, t, up_to_date}; use crate::LLVM_TOOLS; use crate::{CLang, Compiler, DependencyType, GitRepo, Mode}; use filetime::FileTime; #[derive(Debug, Copy, Clone, PartialEq, Eq, PartialOrd, Ord, Hash)] pub struct Std { pub target: TargetSelection, pub compiler: Compiler, /// Whether to build only a subset of crates in the standard library. /// /// This shouldn't be used from other steps; see the comment on [`Rustc`]. crates: Interned>, /// When using download-rustc, we need to use a new build of `std` for running unit tests of Std itself, /// but we need to use the downloaded copy of std for linking to rustdoc. Allow this to be overriden by `builder.ensure` from other steps. force_recompile: bool, } impl Std { pub fn new(compiler: Compiler, target: TargetSelection) -> Self { Self { target, compiler, crates: Default::default(), force_recompile: false } } pub fn force_recompile(compiler: Compiler, target: TargetSelection) -> Self { Self { target, compiler, crates: Default::default(), force_recompile: true } } } impl Step for Std { type Output = (); const DEFAULT: bool = true; fn should_run(run: ShouldRun<'_>) -> ShouldRun<'_> { // When downloading stage1, the standard library has already been copied to the sysroot, so // there's no need to rebuild it. let builder = run.builder; run.crate_or_deps("sysroot") .path("library") .lazy_default_condition(Box::new(|| !builder.download_rustc())) } fn make_run(run: RunConfig<'_>) { // If the paths include "library", build the entire standard library. let has_alias = run.paths.iter().any(|set| set.assert_single_path().path.ends_with("library")); let crates = if has_alias { Default::default() } else { run.cargo_crates_in_set() }; run.builder.ensure(Std { compiler: run.builder.compiler(run.builder.top_stage, run.build_triple()), target: run.target, crates, force_recompile: false, }); } /// Builds the standard library. /// /// This will build the standard library for a particular stage of the build /// using the `compiler` targeting the `target` architecture. The artifacts /// created will also be linked into the sysroot directory. fn run(self, builder: &Builder<'_>) { let target = self.target; let compiler = self.compiler; // When using `download-rustc`, we already have artifacts for the host available. Don't // recompile them. if builder.download_rustc() && target == builder.build.build // NOTE: the beta compiler may generate different artifacts than the downloaded compiler, so // its artifacts can't be reused. && compiler.stage != 0 // This check is specific to testing std itself; see `test::Std` for more details. && !self.force_recompile { cp_rustc_component_to_ci_sysroot( builder, compiler, builder.config.ci_rust_std_contents(), ); return; } if builder.config.keep_stage.contains(&compiler.stage) || builder.config.keep_stage_std.contains(&compiler.stage) { builder.info("Warning: Using a potentially old libstd. This may not behave well."); copy_third_party_objects(builder, &compiler, target); copy_self_contained_objects(builder, &compiler, target); builder.ensure(StdLink::from_std(self, compiler)); return; } builder.update_submodule(&Path::new("library").join("stdarch")); // Profiler information requires LLVM's compiler-rt if builder.config.profiler { builder.update_submodule(&Path::new("src/llvm-project")); } let mut target_deps = builder.ensure(StartupObjects { compiler, target }); let compiler_to_use = builder.compiler_for(compiler.stage, compiler.host, target); if compiler_to_use != compiler { builder.ensure(Std::new(compiler_to_use, target)); let msg = if compiler_to_use.host == target { format!( "Uplifting library (stage{} -> stage{})", compiler_to_use.stage, compiler.stage ) } else { format!( "Uplifting library (stage{}:{} -> stage{}:{})", compiler_to_use.stage, compiler_to_use.host, compiler.stage, target ) }; builder.info(&msg); // Even if we're not building std this stage, the new sysroot must // still contain the third party objects needed by various targets. copy_third_party_objects(builder, &compiler, target); copy_self_contained_objects(builder, &compiler, target); builder.ensure(StdLink::from_std(self, compiler_to_use)); return; } target_deps.extend(copy_third_party_objects(builder, &compiler, target)); target_deps.extend(copy_self_contained_objects(builder, &compiler, target)); let mut cargo = builder.cargo(compiler, Mode::Std, SourceType::InTree, target, "build"); std_cargo(builder, target, compiler.stage, &mut cargo); for krate in &*self.crates { cargo.arg("-p").arg(krate); } // See src/bootstrap/synthetic_targets.rs if target.is_synthetic() { cargo.env("RUSTC_BOOTSTRAP_SYNTHETIC_TARGET", "1"); } let _guard = builder.msg( Kind::Build, compiler.stage, format_args!("library artifacts{}", crate_description(&self.crates)), compiler.host, target, ); run_cargo( builder, cargo, vec![], &libstd_stamp(builder, compiler, target), target_deps, false, false, ); builder.ensure(StdLink::from_std( self, builder.compiler(compiler.stage, builder.config.build), )); } } fn copy_and_stamp( builder: &Builder<'_>, libdir: &Path, sourcedir: &Path, name: &str, target_deps: &mut Vec<(PathBuf, DependencyType)>, dependency_type: DependencyType, ) { let target = libdir.join(name); builder.copy(&sourcedir.join(name), &target); target_deps.push((target, dependency_type)); } fn copy_llvm_libunwind(builder: &Builder<'_>, target: TargetSelection, libdir: &Path) -> PathBuf { let libunwind_path = builder.ensure(llvm::Libunwind { target }); let libunwind_source = libunwind_path.join("libunwind.a"); let libunwind_target = libdir.join("libunwind.a"); builder.copy(&libunwind_source, &libunwind_target); libunwind_target } /// Copies third party objects needed by various targets. fn copy_third_party_objects( builder: &Builder<'_>, compiler: &Compiler, target: TargetSelection, ) -> Vec<(PathBuf, DependencyType)> { let mut target_deps = vec![]; if builder.config.sanitizers_enabled(target) && compiler.stage != 0 { // The sanitizers are only copied in stage1 or above, // to avoid creating dependency on LLVM. target_deps.extend( copy_sanitizers(builder, &compiler, target) .into_iter() .map(|d| (d, DependencyType::Target)), ); } if target == "x86_64-fortanix-unknown-sgx" || builder.config.llvm_libunwind(target) == LlvmLibunwind::InTree && (target.contains("linux") || target.contains("fuchsia")) { let libunwind_path = copy_llvm_libunwind(builder, target, &builder.sysroot_libdir(*compiler, target)); target_deps.push((libunwind_path, DependencyType::Target)); } target_deps } /// Copies third party objects needed by various targets for self-contained linkage. fn copy_self_contained_objects( builder: &Builder<'_>, compiler: &Compiler, target: TargetSelection, ) -> Vec<(PathBuf, DependencyType)> { let libdir_self_contained = builder.sysroot_libdir(*compiler, target).join("self-contained"); t!(fs::create_dir_all(&libdir_self_contained)); let mut target_deps = vec![]; // Copies the libc and CRT objects. // // rustc historically provides a more self-contained installation for musl targets // not requiring the presence of a native musl toolchain. For example, it can fall back // to using gcc from a glibc-targeting toolchain for linking. // To do that we have to distribute musl startup objects as a part of Rust toolchain // and link with them manually in the self-contained mode. if target.contains("musl") && !target.contains("unikraft") { let srcdir = builder.musl_libdir(target).unwrap_or_else(|| { panic!("Target {:?} does not have a \"musl-libdir\" key", target.triple) }); for &obj in &["libc.a", "crt1.o", "Scrt1.o", "rcrt1.o", "crti.o", "crtn.o"] { copy_and_stamp( builder, &libdir_self_contained, &srcdir, obj, &mut target_deps, DependencyType::TargetSelfContained, ); } let crt_path = builder.ensure(llvm::CrtBeginEnd { target }); for &obj in &["crtbegin.o", "crtbeginS.o", "crtend.o", "crtendS.o"] { let src = crt_path.join(obj); let target = libdir_self_contained.join(obj); builder.copy(&src, &target); target_deps.push((target, DependencyType::TargetSelfContained)); } if !target.starts_with("s390x") { let libunwind_path = copy_llvm_libunwind(builder, target, &libdir_self_contained); target_deps.push((libunwind_path, DependencyType::TargetSelfContained)); } } else if target.contains("-wasi") { let srcdir = builder .wasi_root(target) .unwrap_or_else(|| { panic!("Target {:?} does not have a \"wasi-root\" key", target.triple) }) .join("lib") .join(target.to_string().replace("-preview1", "")); for &obj in &["libc.a", "crt1-command.o", "crt1-reactor.o"] { copy_and_stamp( builder, &libdir_self_contained, &srcdir, obj, &mut target_deps, DependencyType::TargetSelfContained, ); } } else if target.ends_with("windows-gnu") { for obj in ["crt2.o", "dllcrt2.o"].iter() { let src = compiler_file(builder, &builder.cc(target), target, CLang::C, obj); let target = libdir_self_contained.join(obj); builder.copy(&src, &target); target_deps.push((target, DependencyType::TargetSelfContained)); } } target_deps } /// Configure cargo to compile the standard library, adding appropriate env vars /// and such. pub fn std_cargo(builder: &Builder<'_>, target: TargetSelection, stage: u32, cargo: &mut Cargo) { if let Some(target) = env::var_os("MACOSX_STD_DEPLOYMENT_TARGET") { cargo.env("MACOSX_DEPLOYMENT_TARGET", target); } if let Some(path) = builder.config.profiler_path(target) { cargo.env("LLVM_PROFILER_RT_LIB", path); } // Determine if we're going to compile in optimized C intrinsics to // the `compiler-builtins` crate. These intrinsics live in LLVM's // `compiler-rt` repository, but our `src/llvm-project` submodule isn't // always checked out, so we need to conditionally look for this. (e.g. if // an external LLVM is used we skip the LLVM submodule checkout). // // Note that this shouldn't affect the correctness of `compiler-builtins`, // but only its speed. Some intrinsics in C haven't been translated to Rust // yet but that's pretty rare. Other intrinsics have optimized // implementations in C which have only had slower versions ported to Rust, // so we favor the C version where we can, but it's not critical. // // If `compiler-rt` is available ensure that the `c` feature of the // `compiler-builtins` crate is enabled and it's configured to learn where // `compiler-rt` is located. let compiler_builtins_root = builder.src.join("src/llvm-project/compiler-rt"); let compiler_builtins_c_feature = if compiler_builtins_root.exists() { // Note that `libprofiler_builtins/build.rs` also computes this so if // you're changing something here please also change that. cargo.env("RUST_COMPILER_RT_ROOT", &compiler_builtins_root); " compiler-builtins-c" } else { "" }; // `libtest` uses this to know whether or not to support // `-Zunstable-options`. if !builder.unstable_features() { cargo.env("CFG_DISABLE_UNSTABLE_FEATURES", "1"); } let mut features = String::new(); // Cranelift doesn't support `asm`. if stage != 0 && builder.config.default_codegen_backend().unwrap_or_default() == "cranelift" { features += " compiler-builtins-no-asm"; } if builder.no_std(target) == Some(true) { features += " compiler-builtins-mem"; if !target.starts_with("bpf") { features.push_str(compiler_builtins_c_feature); } // for no-std targets we only compile a few no_std crates cargo .args(&["-p", "alloc"]) .arg("--manifest-path") .arg(builder.src.join("library/alloc/Cargo.toml")) .arg("--features") .arg(features); } else { features += &builder.std_features(target); features.push_str(compiler_builtins_c_feature); cargo .arg("--features") .arg(features) .arg("--manifest-path") .arg(builder.src.join("library/sysroot/Cargo.toml")); // Help the libc crate compile by assisting it in finding various // sysroot native libraries. if target.contains("musl") { if let Some(p) = builder.musl_libdir(target) { let root = format!("native={}", p.to_str().unwrap()); cargo.rustflag("-L").rustflag(&root); } } if target.contains("-wasi") { if let Some(p) = builder.wasi_root(target) { let root = format!( "native={}/lib/{}", p.to_str().unwrap(), target.to_string().replace("-preview1", "") ); cargo.rustflag("-L").rustflag(&root); } } } // By default, rustc uses `-Cembed-bitcode=yes`, and Cargo overrides that // with `-Cembed-bitcode=no` for non-LTO builds. However, libstd must be // built with bitcode so that the produced rlibs can be used for both LTO // builds (which use bitcode) and non-LTO builds (which use object code). // So we override the override here! // // But we don't bother for the stage 0 compiler because it's never used // with LTO. if stage >= 1 { cargo.rustflag("-Cembed-bitcode=yes"); } if builder.config.rust_lto == RustcLto::Off { cargo.rustflag("-Clto=off"); } // By default, rustc does not include unwind tables unless they are required // for a particular target. They are not required by RISC-V targets, but // compiling the standard library with them means that users can get // backtraces without having to recompile the standard library themselves. // // This choice was discussed in https://github.com/rust-lang/rust/pull/69890 if target.contains("riscv") { cargo.rustflag("-Cforce-unwind-tables=yes"); } let html_root = format!("-Zcrate-attr=doc(html_root_url=\"{}/\")", builder.doc_rust_lang_org_channel(),); cargo.rustflag(&html_root); cargo.rustdocflag(&html_root); cargo.rustdocflag("-Zcrate-attr=warn(rust_2018_idioms)"); } #[derive(Debug, Copy, Clone, PartialEq, Eq, Hash)] struct StdLink { pub compiler: Compiler, pub target_compiler: Compiler, pub target: TargetSelection, /// Not actually used; only present to make sure the cache invalidation is correct. crates: Interned>, /// See [`Std::force_recompile`]. force_recompile: bool, } impl StdLink { fn from_std(std: Std, host_compiler: Compiler) -> Self { Self { compiler: host_compiler, target_compiler: std.compiler, target: std.target, crates: std.crates, force_recompile: std.force_recompile, } } } impl Step for StdLink { type Output = (); fn should_run(run: ShouldRun<'_>) -> ShouldRun<'_> { run.never() } /// Link all libstd rlibs/dylibs into the sysroot location. /// /// Links those artifacts generated by `compiler` to the `stage` compiler's /// sysroot for the specified `host` and `target`. /// /// Note that this assumes that `compiler` has already generated the libstd /// libraries for `target`, and this method will find them in the relevant /// output directory. fn run(self, builder: &Builder<'_>) { let compiler = self.compiler; let target_compiler = self.target_compiler; let target = self.target; // NOTE: intentionally does *not* check `target == builder.build` to avoid having to add the same check in `test::Crate`. let (libdir, hostdir) = if self.force_recompile && builder.download_rustc() { // NOTE: copies part of `sysroot_libdir` to avoid having to add a new `force_recompile` argument there too let lib = builder.sysroot_libdir_relative(self.compiler); let sysroot = builder.ensure(crate::compile::Sysroot { compiler: self.compiler, force_recompile: self.force_recompile, }); let libdir = sysroot.join(lib).join("rustlib").join(target.triple).join("lib"); let hostdir = sysroot.join(lib).join("rustlib").join(compiler.host.triple).join("lib"); (INTERNER.intern_path(libdir), INTERNER.intern_path(hostdir)) } else { let libdir = builder.sysroot_libdir(target_compiler, target); let hostdir = builder.sysroot_libdir(target_compiler, compiler.host); (libdir, hostdir) }; add_to_sysroot(builder, &libdir, &hostdir, &libstd_stamp(builder, compiler, target)); // Special case for stage0, to make `rustup toolchain link` and `x dist --stage 0` // work for stage0-sysroot. We only do this if the stage0 compiler comes from beta, // and is not set to a custom path. if compiler.stage == 0 && builder .build .config .initial_rustc .starts_with(builder.out.join(&compiler.host.triple).join("stage0/bin")) { // Copy bin files from stage0/bin to stage0-sysroot/bin let sysroot = builder.out.join(&compiler.host.triple).join("stage0-sysroot"); let host = compiler.host.triple; let stage0_bin_dir = builder.out.join(&host).join("stage0/bin"); let sysroot_bin_dir = sysroot.join("bin"); t!(fs::create_dir_all(&sysroot_bin_dir)); builder.cp_r(&stage0_bin_dir, &sysroot_bin_dir); // Copy all *.so files from stage0/lib to stage0-sysroot/lib let stage0_lib_dir = builder.out.join(&host).join("stage0/lib"); if let Ok(files) = fs::read_dir(&stage0_lib_dir) { for file in files { let file = t!(file); let path = file.path(); if path.is_file() && is_dylib(&file.file_name().into_string().unwrap()) { builder.copy(&path, &sysroot.join("lib").join(path.file_name().unwrap())); } } } // Copy codegen-backends from stage0 let sysroot_codegen_backends = builder.sysroot_codegen_backends(compiler); t!(fs::create_dir_all(&sysroot_codegen_backends)); let stage0_codegen_backends = builder .out .join(&host) .join("stage0/lib/rustlib") .join(&host) .join("codegen-backends"); builder.cp_r(&stage0_codegen_backends, &sysroot_codegen_backends); } } } /// Copies sanitizer runtime libraries into target libdir. fn copy_sanitizers( builder: &Builder<'_>, compiler: &Compiler, target: TargetSelection, ) -> Vec { let runtimes: Vec = builder.ensure(llvm::Sanitizers { target }); if builder.config.dry_run() { return Vec::new(); } let mut target_deps = Vec::new(); let libdir = builder.sysroot_libdir(*compiler, target); for runtime in &runtimes { let dst = libdir.join(&runtime.name); builder.copy(&runtime.path, &dst); if target == "x86_64-apple-darwin" || target == "aarch64-apple-darwin" || target == "aarch64-apple-ios" || target == "aarch64-apple-ios-sim" || target == "x86_64-apple-ios" { // Update the library’s install name to reflect that it has been renamed. apple_darwin_update_library_name(&dst, &format!("@rpath/{}", &runtime.name)); // Upon renaming the install name, the code signature of the file will invalidate, // so we will sign it again. apple_darwin_sign_file(&dst); } target_deps.push(dst); } target_deps } fn apple_darwin_update_library_name(library_path: &Path, new_name: &str) { let status = Command::new("install_name_tool") .arg("-id") .arg(new_name) .arg(library_path) .status() .expect("failed to execute `install_name_tool`"); assert!(status.success()); } fn apple_darwin_sign_file(file_path: &Path) { let status = Command::new("codesign") .arg("-f") // Force to rewrite the existing signature .arg("-s") .arg("-") .arg(file_path) .status() .expect("failed to execute `codesign`"); assert!(status.success()); } #[derive(Debug, Copy, Clone, PartialEq, Eq, Hash)] pub struct StartupObjects { pub compiler: Compiler, pub target: TargetSelection, } impl Step for StartupObjects { type Output = Vec<(PathBuf, DependencyType)>; fn should_run(run: ShouldRun<'_>) -> ShouldRun<'_> { run.path("library/rtstartup") } fn make_run(run: RunConfig<'_>) { run.builder.ensure(StartupObjects { compiler: run.builder.compiler(run.builder.top_stage, run.build_triple()), target: run.target, }); } /// Builds and prepare startup objects like rsbegin.o and rsend.o /// /// These are primarily used on Windows right now for linking executables/dlls. /// They don't require any library support as they're just plain old object /// files, so we just use the nightly snapshot compiler to always build them (as /// no other compilers are guaranteed to be available). fn run(self, builder: &Builder<'_>) -> Vec<(PathBuf, DependencyType)> { let for_compiler = self.compiler; let target = self.target; if !target.ends_with("windows-gnu") { return vec![]; } let mut target_deps = vec![]; let src_dir = &builder.src.join("library").join("rtstartup"); let dst_dir = &builder.native_dir(target).join("rtstartup"); let sysroot_dir = &builder.sysroot_libdir(for_compiler, target); t!(fs::create_dir_all(dst_dir)); for file in &["rsbegin", "rsend"] { let src_file = &src_dir.join(file.to_string() + ".rs"); let dst_file = &dst_dir.join(file.to_string() + ".o"); if !up_to_date(src_file, dst_file) { let mut cmd = Command::new(&builder.initial_rustc); cmd.env("RUSTC_BOOTSTRAP", "1"); if !builder.local_rebuild { // a local_rebuild compiler already has stage1 features cmd.arg("--cfg").arg("bootstrap"); } builder.run( cmd.arg("--target") .arg(target.rustc_target_arg()) .arg("--emit=obj") .arg("-o") .arg(dst_file) .arg(src_file), ); } let target = sysroot_dir.join((*file).to_string() + ".o"); builder.copy(dst_file, &target); target_deps.push((target, DependencyType::Target)); } target_deps } } fn cp_rustc_component_to_ci_sysroot( builder: &Builder<'_>, compiler: Compiler, contents: Vec, ) { let sysroot = builder.ensure(Sysroot { compiler, force_recompile: false }); let ci_rustc_dir = builder.config.ci_rustc_dir(); for file in contents { let src = ci_rustc_dir.join(&file); let dst = sysroot.join(file); if src.is_dir() { t!(fs::create_dir_all(dst)); } else { builder.copy(&src, &dst); } } } #[derive(Debug, PartialOrd, Ord, Copy, Clone, PartialEq, Eq, Hash)] pub struct Rustc { pub target: TargetSelection, pub compiler: Compiler, /// Whether to build a subset of crates, rather than the whole compiler. /// /// This should only be requested by the user, not used within rustbuild itself. /// Using it within rustbuild can lead to confusing situation where lints are replayed /// in two different steps. crates: Interned>, } impl Rustc { pub fn new(compiler: Compiler, target: TargetSelection) -> Self { Self { target, compiler, crates: Default::default() } } } impl Step for Rustc { type Output = (); const ONLY_HOSTS: bool = true; const DEFAULT: bool = false; fn should_run(run: ShouldRun<'_>) -> ShouldRun<'_> { let mut crates = run.builder.in_tree_crates("rustc-main", None); for (i, krate) in crates.iter().enumerate() { // We can't allow `build rustc` as an alias for this Step, because that's reserved by `Assemble`. // Ideally Assemble would use `build compiler` instead, but that seems too confusing to be worth the breaking change. if krate.name == "rustc-main" { crates.swap_remove(i); break; } } run.crates(crates) } fn make_run(run: RunConfig<'_>) { let crates = run.cargo_crates_in_set(); run.builder.ensure(Rustc { compiler: run.builder.compiler(run.builder.top_stage, run.build_triple()), target: run.target, crates, }); } /// Builds the compiler. /// /// This will build the compiler for a particular stage of the build using /// the `compiler` targeting the `target` architecture. The artifacts /// created will also be linked into the sysroot directory. fn run(self, builder: &Builder<'_>) { let compiler = self.compiler; let target = self.target; // NOTE: the ABI of the beta compiler is different from the ABI of the downloaded compiler, // so its artifacts can't be reused. if builder.download_rustc() && compiler.stage != 0 { // Copy the existing artifacts instead of rebuilding them. // NOTE: this path is only taken for tools linking to rustc-dev (including ui-fulldeps tests). cp_rustc_component_to_ci_sysroot( builder, compiler, builder.config.ci_rustc_dev_contents(), ); return; } builder.ensure(Std::new(compiler, target)); if builder.config.keep_stage.contains(&compiler.stage) { builder.info("Warning: Using a potentially old librustc. This may not behave well."); builder.info("Warning: Use `--keep-stage-std` if you want to rebuild the compiler when it changes"); builder.ensure(RustcLink::from_rustc(self, compiler)); return; } let compiler_to_use = builder.compiler_for(compiler.stage, compiler.host, target); if compiler_to_use != compiler { builder.ensure(Rustc::new(compiler_to_use, target)); let msg = if compiler_to_use.host == target { format!( "Uplifting rustc (stage{} -> stage{})", compiler_to_use.stage, compiler.stage + 1 ) } else { format!( "Uplifting rustc (stage{}:{} -> stage{}:{})", compiler_to_use.stage, compiler_to_use.host, compiler.stage + 1, target ) }; builder.info(&msg); builder.ensure(RustcLink::from_rustc(self, compiler_to_use)); return; } // Ensure that build scripts and proc macros have a std / libproc_macro to link against. builder.ensure(Std::new( builder.compiler(self.compiler.stage, builder.config.build), builder.config.build, )); let mut cargo = builder.cargo(compiler, Mode::Rustc, SourceType::InTree, target, "build"); rustc_cargo(builder, &mut cargo, target, compiler.stage); if builder.config.rust_profile_use.is_some() && builder.config.rust_profile_generate.is_some() { panic!("Cannot use and generate PGO profiles at the same time"); } // With LLD, we can use ICF (identical code folding) to reduce the executable size // of librustc_driver/rustc and to improve i-cache utilization. // // -Wl,[link options] doesn't work on MSVC. However, /OPT:ICF (technically /OPT:REF,ICF) // is already on by default in MSVC optimized builds, which is interpreted as --icf=all: // https://github.com/llvm/llvm-project/blob/3329cec2f79185bafd678f310fafadba2a8c76d2/lld/COFF/Driver.cpp#L1746 // https://github.com/rust-lang/rust/blob/f22819bcce4abaff7d1246a56eec493418f9f4ee/compiler/rustc_codegen_ssa/src/back/linker.rs#L827 if builder.config.use_lld && !compiler.host.contains("msvc") { cargo.rustflag("-Clink-args=-Wl,--icf=all"); } let is_collecting = if let Some(path) = &builder.config.rust_profile_generate { if compiler.stage == 1 { cargo.rustflag(&format!("-Cprofile-generate={path}")); // Apparently necessary to avoid overflowing the counters during // a Cargo build profile cargo.rustflag("-Cllvm-args=-vp-counters-per-site=4"); true } else { false } } else if let Some(path) = &builder.config.rust_profile_use { if compiler.stage == 1 { cargo.rustflag(&format!("-Cprofile-use={path}")); cargo.rustflag("-Cllvm-args=-pgo-warn-missing-function"); true } else { false } } else { false }; if is_collecting { // Ensure paths to Rust sources are relative, not absolute. cargo.rustflag(&format!( "-Cllvm-args=-static-func-strip-dirname-prefix={}", builder.config.src.components().count() )); } // We currently don't support cross-crate LTO in stage0. This also isn't hugely necessary // and may just be a time sink. if compiler.stage != 0 { match builder.config.rust_lto { RustcLto::Thin | RustcLto::Fat => { // Since using LTO for optimizing dylibs is currently experimental, // we need to pass -Zdylib-lto. cargo.rustflag("-Zdylib-lto"); // Cargo by default passes `-Cembed-bitcode=no` and doesn't pass `-Clto` when // compiling dylibs (and their dependencies), even when LTO is enabled for the // crate. Therefore, we need to override `-Clto` and `-Cembed-bitcode` here. let lto_type = match builder.config.rust_lto { RustcLto::Thin => "thin", RustcLto::Fat => "fat", _ => unreachable!(), }; cargo.rustflag(&format!("-Clto={lto_type}")); cargo.rustflag("-Cembed-bitcode=yes"); } RustcLto::ThinLocal => { /* Do nothing, this is the default */ } RustcLto::Off => { cargo.rustflag("-Clto=off"); } } } else { if builder.config.rust_lto == RustcLto::Off { cargo.rustflag("-Clto=off"); } } for krate in &*self.crates { cargo.arg("-p").arg(krate); } let _guard = builder.msg_sysroot_tool( Kind::Build, compiler.stage, format_args!("compiler artifacts{}", crate_description(&self.crates)), compiler.host, target, ); let stamp = librustc_stamp(builder, compiler, target); run_cargo( builder, cargo, vec![], &stamp, vec![], false, true, // Only ship rustc_driver.so and .rmeta files, not all intermediate .rlib files. ); // When building `librustc_driver.so` (like `libLLVM.so`) on linux, it can contain // unexpected debuginfo from dependencies, for example from the C++ standard library used in // our LLVM wrapper. Unless we're explicitly requesting `librustc_driver` to be built with // debuginfo (via the debuginfo level of the executables using it): strip this debuginfo // away after the fact. if builder.config.rust_debuginfo_level_rustc == DebuginfoLevel::None && builder.config.rust_debuginfo_level_tools == DebuginfoLevel::None { let target_root_dir = stamp.parent().unwrap(); let rustc_driver = target_root_dir.join("librustc_driver.so"); strip_debug(builder, target, &rustc_driver); } builder.ensure(RustcLink::from_rustc( self, builder.compiler(compiler.stage, builder.config.build), )); } } pub fn rustc_cargo(builder: &Builder<'_>, cargo: &mut Cargo, target: TargetSelection, stage: u32) { cargo .arg("--features") .arg(builder.rustc_features(builder.kind)) .arg("--manifest-path") .arg(builder.src.join("compiler/rustc/Cargo.toml")); cargo.rustdocflag("-Zcrate-attr=warn(rust_2018_idioms)"); rustc_cargo_env(builder, cargo, target, stage); } pub fn rustc_cargo_env( builder: &Builder<'_>, cargo: &mut Cargo, target: TargetSelection, stage: u32, ) { // Set some configuration variables picked up by build scripts and // the compiler alike cargo .env("CFG_RELEASE", builder.rust_release()) .env("CFG_RELEASE_CHANNEL", &builder.config.channel) .env("CFG_VERSION", builder.rust_version()); if let Some(backend) = builder.config.default_codegen_backend() { cargo.env("CFG_DEFAULT_CODEGEN_BACKEND", backend); } let libdir_relative = builder.config.libdir_relative().unwrap_or_else(|| Path::new("lib")); let target_config = builder.config.target_config.get(&target); cargo.env("CFG_LIBDIR_RELATIVE", libdir_relative); if let Some(ref ver_date) = builder.rust_info().commit_date() { cargo.env("CFG_VER_DATE", ver_date); } if let Some(ref ver_hash) = builder.rust_info().sha() { cargo.env("CFG_VER_HASH", ver_hash); } if !builder.unstable_features() { cargo.env("CFG_DISABLE_UNSTABLE_FEATURES", "1"); } // Prefer the current target's own default_linker, else a globally // specified one. if let Some(s) = target_config.and_then(|c| c.default_linker.as_ref()) { cargo.env("CFG_DEFAULT_LINKER", s); } else if let Some(ref s) = builder.config.rustc_default_linker { cargo.env("CFG_DEFAULT_LINKER", s); } if builder.config.rustc_parallel { // keep in sync with `bootstrap/lib.rs:Build::rustc_features` // `cfg` option for rustc, `features` option for cargo, for conditional compilation cargo.rustflag("--cfg=parallel_compiler"); cargo.rustdocflag("--cfg=parallel_compiler"); } if builder.config.rust_verify_llvm_ir { cargo.env("RUSTC_VERIFY_LLVM_IR", "1"); } // Note that this is disabled if LLVM itself is disabled or we're in a check // build. If we are in a check build we still go ahead here presuming we've // detected that LLVM is already built and good to go which helps prevent // busting caches (e.g. like #71152). if builder.config.llvm_enabled() { let building_is_expensive = crate::llvm::prebuilt_llvm_config(builder, target).is_err(); // `top_stage == stage` might be false for `check --stage 1`, if we are building the stage 1 compiler let can_skip_build = builder.kind == Kind::Check && builder.top_stage == stage; let should_skip_build = building_is_expensive && can_skip_build; if !should_skip_build { rustc_llvm_env(builder, cargo, target) } } } /// Pass down configuration from the LLVM build into the build of /// rustc_llvm and rustc_codegen_llvm. fn rustc_llvm_env(builder: &Builder<'_>, cargo: &mut Cargo, target: TargetSelection) { let target_config = builder.config.target_config.get(&target); if builder.is_rust_llvm(target) { cargo.env("LLVM_RUSTLLVM", "1"); } let llvm::LlvmResult { llvm_config, .. } = builder.ensure(llvm::Llvm { target }); cargo.env("LLVM_CONFIG", &llvm_config); if let Some(s) = target_config.and_then(|c| c.llvm_config.as_ref()) { cargo.env("CFG_LLVM_ROOT", s); } // Some LLVM linker flags (-L and -l) may be needed to link `rustc_llvm`. Its build script // expects these to be passed via the `LLVM_LINKER_FLAGS` env variable, separated by // whitespace. // // For example: // - on windows, when `clang-cl` is used with instrumentation, we need to manually add // clang's runtime library resource directory so that the profiler runtime library can be // found. This is to avoid the linker errors about undefined references to // `__llvm_profile_instrument_memop` when linking `rustc_driver`. let mut llvm_linker_flags = String::new(); if builder.config.llvm_profile_generate && target.contains("msvc") { if let Some(ref clang_cl_path) = builder.config.llvm_clang_cl { // Add clang's runtime library directory to the search path let clang_rt_dir = get_clang_cl_resource_dir(clang_cl_path); llvm_linker_flags.push_str(&format!("-L{}", clang_rt_dir.display())); } } // The config can also specify its own llvm linker flags. if let Some(ref s) = builder.config.llvm_ldflags { if !llvm_linker_flags.is_empty() { llvm_linker_flags.push_str(" "); } llvm_linker_flags.push_str(s); } // Set the linker flags via the env var that `rustc_llvm`'s build script will read. if !llvm_linker_flags.is_empty() { cargo.env("LLVM_LINKER_FLAGS", llvm_linker_flags); } // Building with a static libstdc++ is only supported on linux right now, // not for MSVC or macOS if builder.config.llvm_static_stdcpp && !target.contains("freebsd") && !target.contains("msvc") && !target.contains("apple") && !target.contains("solaris") { let file = compiler_file( builder, &builder.cxx(target).unwrap(), target, CLang::Cxx, "libstdc++.a", ); cargo.env("LLVM_STATIC_STDCPP", file); } if builder.llvm_link_shared() { cargo.env("LLVM_LINK_SHARED", "1"); } if builder.config.llvm_use_libcxx { cargo.env("LLVM_USE_LIBCXX", "1"); } if builder.config.llvm_optimize && !builder.config.llvm_release_debuginfo { cargo.env("LLVM_NDEBUG", "1"); } } #[derive(Debug, Copy, Clone, PartialEq, Eq, Hash)] struct RustcLink { pub compiler: Compiler, pub target_compiler: Compiler, pub target: TargetSelection, /// Not actually used; only present to make sure the cache invalidation is correct. crates: Interned>, } impl RustcLink { fn from_rustc(rustc: Rustc, host_compiler: Compiler) -> Self { Self { compiler: host_compiler, target_compiler: rustc.compiler, target: rustc.target, crates: rustc.crates, } } } impl Step for RustcLink { type Output = (); fn should_run(run: ShouldRun<'_>) -> ShouldRun<'_> { run.never() } /// Same as `std_link`, only for librustc fn run(self, builder: &Builder<'_>) { let compiler = self.compiler; let target_compiler = self.target_compiler; let target = self.target; add_to_sysroot( builder, &builder.sysroot_libdir(target_compiler, target), &builder.sysroot_libdir(target_compiler, compiler.host), &librustc_stamp(builder, compiler, target), ); } } #[derive(Debug, Copy, Clone, PartialEq, Eq, Hash)] pub struct CodegenBackend { pub target: TargetSelection, pub compiler: Compiler, pub backend: Interned, } fn needs_codegen_config(run: &RunConfig<'_>) -> bool { let mut needs_codegen_cfg = false; for path_set in &run.paths { needs_codegen_cfg = match path_set { PathSet::Set(set) => set.iter().any(|p| is_codegen_cfg_needed(p, run)), PathSet::Suite(suite) => is_codegen_cfg_needed(&suite, run), } } needs_codegen_cfg } pub(crate) const CODEGEN_BACKEND_PREFIX: &str = "rustc_codegen_"; fn is_codegen_cfg_needed(path: &TaskPath, run: &RunConfig<'_>) -> bool { if path.path.to_str().unwrap().contains(&CODEGEN_BACKEND_PREFIX) { let mut needs_codegen_backend_config = true; for &backend in &run.builder.config.rust_codegen_backends { if path .path .to_str() .unwrap() .ends_with(&(CODEGEN_BACKEND_PREFIX.to_owned() + &backend)) { needs_codegen_backend_config = false; } } if needs_codegen_backend_config { run.builder.info( "Warning: no codegen-backends config matched the requested path to build a codegen backend. \ Help: add backend to codegen-backends in config.toml.", ); return true; } } return false; } impl Step for CodegenBackend { type Output = (); const ONLY_HOSTS: bool = true; // Only the backends specified in the `codegen-backends` entry of `config.toml` are built. const DEFAULT: bool = true; fn should_run(run: ShouldRun<'_>) -> ShouldRun<'_> { run.paths(&["compiler/rustc_codegen_cranelift", "compiler/rustc_codegen_gcc"]) } fn make_run(run: RunConfig<'_>) { if needs_codegen_config(&run) { return; } for &backend in &run.builder.config.rust_codegen_backends { if backend == "llvm" { continue; // Already built as part of rustc } run.builder.ensure(CodegenBackend { target: run.target, compiler: run.builder.compiler(run.builder.top_stage, run.build_triple()), backend, }); } } fn run(self, builder: &Builder<'_>) { let compiler = self.compiler; let target = self.target; let backend = self.backend; builder.ensure(Rustc::new(compiler, target)); if builder.config.keep_stage.contains(&compiler.stage) { builder.info( "Warning: Using a potentially old codegen backend. \ This may not behave well.", ); // Codegen backends are linked separately from this step today, so we don't do // anything here. return; } let compiler_to_use = builder.compiler_for(compiler.stage, compiler.host, target); if compiler_to_use != compiler { builder.ensure(CodegenBackend { compiler: compiler_to_use, target, backend }); return; } let out_dir = builder.cargo_out(compiler, Mode::Codegen, target); let mut cargo = builder.cargo(compiler, Mode::Codegen, SourceType::InTree, target, "build"); cargo .arg("--manifest-path") .arg(builder.src.join(format!("compiler/rustc_codegen_{backend}/Cargo.toml"))); rustc_cargo_env(builder, &mut cargo, target, compiler.stage); let tmp_stamp = out_dir.join(".tmp.stamp"); let _guard = builder.msg_build(compiler, format_args!("codegen backend {backend}"), target); let files = run_cargo(builder, cargo, vec![], &tmp_stamp, vec![], false, false); if builder.config.dry_run() { return; } let mut files = files.into_iter().filter(|f| { let filename = f.file_name().unwrap().to_str().unwrap(); is_dylib(filename) && filename.contains("rustc_codegen_") }); let codegen_backend = match files.next() { Some(f) => f, None => panic!("no dylibs built for codegen backend?"), }; if let Some(f) = files.next() { panic!( "codegen backend built two dylibs:\n{}\n{}", codegen_backend.display(), f.display() ); } let stamp = codegen_backend_stamp(builder, compiler, target, backend); let codegen_backend = codegen_backend.to_str().unwrap(); t!(fs::write(&stamp, &codegen_backend)); } } /// Creates the `codegen-backends` folder for a compiler that's about to be /// assembled as a complete compiler. /// /// This will take the codegen artifacts produced by `compiler` and link them /// into an appropriate location for `target_compiler` to be a functional /// compiler. fn copy_codegen_backends_to_sysroot( builder: &Builder<'_>, compiler: Compiler, target_compiler: Compiler, ) { let target = target_compiler.host; // Note that this step is different than all the other `*Link` steps in // that it's not assembling a bunch of libraries but rather is primarily // moving the codegen backend into place. The codegen backend of rustc is // not linked into the main compiler by default but is rather dynamically // selected at runtime for inclusion. // // Here we're looking for the output dylib of the `CodegenBackend` step and // we're copying that into the `codegen-backends` folder. let dst = builder.sysroot_codegen_backends(target_compiler); t!(fs::create_dir_all(&dst), dst); if builder.config.dry_run() { return; } for backend in builder.config.rust_codegen_backends.iter() { if backend == "llvm" { continue; // Already built as part of rustc } let stamp = codegen_backend_stamp(builder, compiler, target, *backend); let dylib = t!(fs::read_to_string(&stamp)); let file = Path::new(&dylib); let filename = file.file_name().unwrap().to_str().unwrap(); // change `librustc_codegen_cranelift-xxxxxx.so` to // `librustc_codegen_cranelift-release.so` let target_filename = { let dash = filename.find('-').unwrap(); let dot = filename.find('.').unwrap(); format!("{}-{}{}", &filename[..dash], builder.rust_release(), &filename[dot..]) }; builder.copy(&file, &dst.join(target_filename)); } } /// Cargo's output path for the standard library in a given stage, compiled /// by a particular compiler for the specified target. pub fn libstd_stamp(builder: &Builder<'_>, compiler: Compiler, target: TargetSelection) -> PathBuf { builder.cargo_out(compiler, Mode::Std, target).join(".libstd.stamp") } /// Cargo's output path for librustc in a given stage, compiled by a particular /// compiler for the specified target. pub fn librustc_stamp( builder: &Builder<'_>, compiler: Compiler, target: TargetSelection, ) -> PathBuf { builder.cargo_out(compiler, Mode::Rustc, target).join(".librustc.stamp") } /// Cargo's output path for librustc_codegen_llvm in a given stage, compiled by a particular /// compiler for the specified target and backend. fn codegen_backend_stamp( builder: &Builder<'_>, compiler: Compiler, target: TargetSelection, backend: Interned, ) -> PathBuf { builder .cargo_out(compiler, Mode::Codegen, target) .join(format!(".librustc_codegen_{backend}.stamp")) } pub fn compiler_file( builder: &Builder<'_>, compiler: &Path, target: TargetSelection, c: CLang, file: &str, ) -> PathBuf { if builder.config.dry_run() { return PathBuf::new(); } let mut cmd = Command::new(compiler); cmd.args(builder.cflags(target, GitRepo::Rustc, c)); cmd.arg(format!("-print-file-name={file}")); let out = output(&mut cmd); PathBuf::from(out.trim()) } #[derive(Debug, Copy, Clone, PartialEq, Eq, Hash)] pub struct Sysroot { pub compiler: Compiler, /// See [`Std::force_recompile`]. force_recompile: bool, } impl Sysroot { pub(crate) fn new(compiler: Compiler) -> Self { Sysroot { compiler, force_recompile: false } } } impl Step for Sysroot { type Output = Interned; fn should_run(run: ShouldRun<'_>) -> ShouldRun<'_> { run.never() } /// Returns the sysroot for the `compiler` specified that *this build system /// generates*. /// /// That is, the sysroot for the stage0 compiler is not what the compiler /// thinks it is by default, but it's the same as the default for stages /// 1-3. fn run(self, builder: &Builder<'_>) -> Interned { let compiler = self.compiler; let host_dir = builder.out.join(&compiler.host.triple); let sysroot_dir = |stage| { if stage == 0 { host_dir.join("stage0-sysroot") } else if self.force_recompile && stage == compiler.stage { host_dir.join(format!("stage{stage}-test-sysroot")) } else if builder.download_rustc() && compiler.stage != builder.top_stage { host_dir.join("ci-rustc-sysroot") } else { host_dir.join(format!("stage{}", stage)) } }; let sysroot = sysroot_dir(compiler.stage); builder.verbose(&format!("Removing sysroot {} to avoid caching bugs", sysroot.display())); let _ = fs::remove_dir_all(&sysroot); t!(fs::create_dir_all(&sysroot)); // In some cases(see https://github.com/rust-lang/rust/issues/109314), when the stage0 // compiler relies on more recent version of LLVM than the beta compiler, it may not // be able to locate the correct LLVM in the sysroot. This situation typically occurs // when we upgrade LLVM version while the beta compiler continues to use an older version. // // Make sure to add the correct version of LLVM into the stage0 sysroot. if compiler.stage == 0 { dist::maybe_install_llvm_target(builder, compiler.host, &sysroot); } // If we're downloading a compiler from CI, we can use the same compiler for all stages other than 0. if builder.download_rustc() && compiler.stage != 0 { assert_eq!( builder.config.build, compiler.host, "Cross-compiling is not yet supported with `download-rustc`", ); // #102002, cleanup old toolchain folders when using download-rustc so people don't use them by accident. for stage in 0..=2 { if stage != compiler.stage { let dir = sysroot_dir(stage); if !dir.ends_with("ci-rustc-sysroot") { let _ = fs::remove_dir_all(dir); } } } // Copy the compiler into the correct sysroot. // NOTE(#108767): We intentionally don't copy `rustc-dev` artifacts until they're requested with `builder.ensure(Rustc)`. // This fixes an issue where we'd have multiple copies of libc in the sysroot with no way to tell which to load. // There are a few quirks of bootstrap that interact to make this reliable: // 1. The order `Step`s are run is hard-coded in `builder.rs` and not configurable. This // avoids e.g. reordering `test::UiFulldeps` before `test::Ui` and causing the latter to // fail because of duplicate metadata. // 2. The sysroot is deleted and recreated between each invocation, so running `x test // ui-fulldeps && x test ui` can't cause failures. let mut filtered_files = Vec::new(); let mut add_filtered_files = |suffix, contents| { for path in contents { let path = Path::new(&path); if path.parent().map_or(false, |parent| parent.ends_with(&suffix)) { filtered_files.push(path.file_name().unwrap().to_owned()); } } }; let suffix = format!("lib/rustlib/{}/lib", compiler.host); add_filtered_files(suffix.as_str(), builder.config.ci_rustc_dev_contents()); // NOTE: we can't copy std eagerly because `stage2-test-sysroot` needs to have only the // newly compiled std, not the downloaded std. add_filtered_files("lib", builder.config.ci_rust_std_contents()); let filtered_extensions = [ OsStr::new("rmeta"), OsStr::new("rlib"), // FIXME: this is wrong when compiler.host != build, but we don't support that today OsStr::new(std::env::consts::DLL_EXTENSION), ]; let ci_rustc_dir = builder.config.ci_rustc_dir(); builder.cp_filtered(&ci_rustc_dir, &sysroot, &|path| { if path.extension().map_or(true, |ext| !filtered_extensions.contains(&ext)) { return true; } if !path.parent().map_or(true, |p| p.ends_with(&suffix)) { return true; } if !filtered_files.iter().all(|f| f != path.file_name().unwrap()) { builder.verbose_than(1, &format!("ignoring {}", path.display())); false } else { true } }); } // Symlink the source root into the same location inside the sysroot, // where `rust-src` component would go (`$sysroot/lib/rustlib/src/rust`), // so that any tools relying on `rust-src` also work for local builds, // and also for translating the virtual `/rustc/$hash` back to the real // directory (for running tests with `rust.remap-debuginfo = true`). let sysroot_lib_rustlib_src = sysroot.join("lib/rustlib/src"); t!(fs::create_dir_all(&sysroot_lib_rustlib_src)); let sysroot_lib_rustlib_src_rust = sysroot_lib_rustlib_src.join("rust"); if let Err(e) = symlink_dir(&builder.config, &builder.src, &sysroot_lib_rustlib_src_rust) { eprintln!( "warning: creating symbolic link `{}` to `{}` failed with {}", sysroot_lib_rustlib_src_rust.display(), builder.src.display(), e, ); if builder.config.rust_remap_debuginfo { eprintln!( "warning: some `tests/ui` tests will fail when lacking `{}`", sysroot_lib_rustlib_src_rust.display(), ); } } // Same for the rustc-src component. let sysroot_lib_rustlib_rustcsrc = sysroot.join("lib/rustlib/rustc-src"); t!(fs::create_dir_all(&sysroot_lib_rustlib_rustcsrc)); let sysroot_lib_rustlib_rustcsrc_rust = sysroot_lib_rustlib_rustcsrc.join("rust"); if let Err(e) = symlink_dir(&builder.config, &builder.src, &sysroot_lib_rustlib_rustcsrc_rust) { eprintln!( "warning: creating symbolic link `{}` to `{}` failed with {}", sysroot_lib_rustlib_rustcsrc_rust.display(), builder.src.display(), e, ); } INTERNER.intern_path(sysroot) } } #[derive(Debug, Copy, PartialOrd, Ord, Clone, PartialEq, Eq, Hash)] pub struct Assemble { /// The compiler which we will produce in this step. Assemble itself will /// take care of ensuring that the necessary prerequisites to do so exist, /// that is, this target can be a stage2 compiler and Assemble will build /// previous stages for you. pub target_compiler: Compiler, } impl Step for Assemble { type Output = Compiler; const ONLY_HOSTS: bool = true; fn should_run(run: ShouldRun<'_>) -> ShouldRun<'_> { run.path("compiler/rustc").path("compiler") } fn make_run(run: RunConfig<'_>) { run.builder.ensure(Assemble { target_compiler: run.builder.compiler(run.builder.top_stage + 1, run.target), }); } /// Prepare a new compiler from the artifacts in `stage` /// /// This will assemble a compiler in `build/$host/stage$stage`. The compiler /// must have been previously produced by the `stage - 1` builder.build /// compiler. fn run(self, builder: &Builder<'_>) -> Compiler { let target_compiler = self.target_compiler; if target_compiler.stage == 0 { assert_eq!( builder.config.build, target_compiler.host, "Cannot obtain compiler for non-native build triple at stage 0" ); // The stage 0 compiler for the build triple is always pre-built. return target_compiler; } // Get the compiler that we'll use to bootstrap ourselves. // // Note that this is where the recursive nature of the bootstrap // happens, as this will request the previous stage's compiler on // downwards to stage 0. // // Also note that we're building a compiler for the host platform. We // only assume that we can run `build` artifacts, which means that to // produce some other architecture compiler we need to start from // `build` to get there. // // FIXME: It may be faster if we build just a stage 1 compiler and then // use that to bootstrap this compiler forward. let build_compiler = builder.compiler(target_compiler.stage - 1, builder.config.build); // If we're downloading a compiler from CI, we can use the same compiler for all stages other than 0. if builder.download_rustc() { let sysroot = builder.ensure(Sysroot { compiler: target_compiler, force_recompile: false }); // Ensure that `libLLVM.so` ends up in the newly created target directory, // so that tools using `rustc_private` can use it. dist::maybe_install_llvm_target(builder, target_compiler.host, &sysroot); // Lower stages use `ci-rustc-sysroot`, not stageN if target_compiler.stage == builder.top_stage { builder.info(&format!("Creating a sysroot for stage{stage} compiler (use `rustup toolchain link 'name' build/host/stage{stage}`)", stage=target_compiler.stage)); } return target_compiler; } // Build the libraries for this compiler to link to (i.e., the libraries // it uses at runtime). NOTE: Crates the target compiler compiles don't // link to these. (FIXME: Is that correct? It seems to be correct most // of the time but I think we do link to these for stage2/bin compilers // when not performing a full bootstrap). builder.ensure(Rustc::new(build_compiler, target_compiler.host)); // FIXME: For now patch over problems noted in #90244 by early returning here, even though // we've not properly assembled the target sysroot. A full fix is pending further investigation, // for now full bootstrap usage is rare enough that this is OK. if target_compiler.stage >= 3 && !builder.config.full_bootstrap { return target_compiler; } for &backend in builder.config.rust_codegen_backends.iter() { if backend == "llvm" { continue; // Already built as part of rustc } builder.ensure(CodegenBackend { compiler: build_compiler, target: target_compiler.host, backend, }); } let lld_install = if builder.config.lld_enabled { Some(builder.ensure(llvm::Lld { target: target_compiler.host })) } else { None }; let stage = target_compiler.stage; let host = target_compiler.host; let (host_info, dir_name) = if build_compiler.host == host { ("".into(), "host".into()) } else { (format!(" ({host})"), host.to_string()) }; // NOTE: "Creating a sysroot" is somewhat inconsistent with our internal terminology, since // sysroots can temporarily be empty until we put the compiler inside. However, // `ensure(Sysroot)` isn't really something that's user facing, so there shouldn't be any // ambiguity. let msg = format!( "Creating a sysroot for stage{stage} compiler{host_info} (use `rustup toolchain link 'name' build/{dir_name}/stage{stage}`)" ); builder.info(&msg); // Link in all dylibs to the libdir let stamp = librustc_stamp(builder, build_compiler, target_compiler.host); let proc_macros = builder .read_stamp_file(&stamp) .into_iter() .filter_map(|(path, dependency_type)| { if dependency_type == DependencyType::Host { Some(path.file_name().unwrap().to_owned().into_string().unwrap()) } else { None } }) .collect::>(); let sysroot = builder.sysroot(target_compiler); let rustc_libdir = builder.rustc_libdir(target_compiler); t!(fs::create_dir_all(&rustc_libdir)); let src_libdir = builder.sysroot_libdir(build_compiler, host); for f in builder.read_dir(&src_libdir) { let filename = f.file_name().into_string().unwrap(); if (is_dylib(&filename) || is_debug_info(&filename)) && !proc_macros.contains(&filename) { builder.copy(&f.path(), &rustc_libdir.join(&filename)); } } copy_codegen_backends_to_sysroot(builder, build_compiler, target_compiler); // We prepend this bin directory to the user PATH when linking Rust binaries. To // avoid shadowing the system LLD we rename the LLD we provide to `rust-lld`. let libdir = builder.sysroot_libdir(target_compiler, target_compiler.host); let libdir_bin = libdir.parent().unwrap().join("bin"); t!(fs::create_dir_all(&libdir_bin)); if let Some(lld_install) = lld_install { let src_exe = exe("lld", target_compiler.host); let dst_exe = exe("rust-lld", target_compiler.host); builder.copy(&lld_install.join("bin").join(&src_exe), &libdir_bin.join(&dst_exe)); // for `-Z gcc-ld=lld` let gcc_ld_dir = libdir_bin.join("gcc-ld"); t!(fs::create_dir(&gcc_ld_dir)); let lld_wrapper_exe = builder.ensure(crate::tool::LldWrapper { compiler: build_compiler, target: target_compiler.host, }); for name in crate::LLD_FILE_NAMES { builder.copy(&lld_wrapper_exe, &gcc_ld_dir.join(exe(name, target_compiler.host))); } } if builder.config.rust_codegen_backends.contains(&INTERNER.intern_str("llvm")) { let llvm::LlvmResult { llvm_config, .. } = builder.ensure(llvm::Llvm { target: target_compiler.host }); if !builder.config.dry_run() { let llvm_bin_dir = output(Command::new(llvm_config).arg("--bindir")); let llvm_bin_dir = Path::new(llvm_bin_dir.trim()); // Since we've already built the LLVM tools, install them to the sysroot. // This is the equivalent of installing the `llvm-tools-preview` component via // rustup, and lets developers use a locally built toolchain to // build projects that expect llvm tools to be present in the sysroot // (e.g. the `bootimage` crate). for tool in LLVM_TOOLS { let tool_exe = exe(tool, target_compiler.host); let src_path = llvm_bin_dir.join(&tool_exe); // When using `download-ci-llvm`, some of the tools // may not exist, so skip trying to copy them. if src_path.exists() { builder.copy(&src_path, &libdir_bin.join(&tool_exe)); } } } } // Ensure that `libLLVM.so` ends up in the newly build compiler directory, // so that it can be found when the newly built `rustc` is run. dist::maybe_install_llvm_runtime(builder, target_compiler.host, &sysroot); dist::maybe_install_llvm_target(builder, target_compiler.host, &sysroot); // Link the compiler binary itself into place let out_dir = builder.cargo_out(build_compiler, Mode::Rustc, host); let rustc = out_dir.join(exe("rustc-main", host)); let bindir = sysroot.join("bin"); t!(fs::create_dir_all(&bindir)); let compiler = builder.rustc(target_compiler); builder.copy(&rustc, &compiler); target_compiler } } /// Link some files into a rustc sysroot. /// /// For a particular stage this will link the file listed in `stamp` into the /// `sysroot_dst` provided. pub fn add_to_sysroot( builder: &Builder<'_>, sysroot_dst: &Path, sysroot_host_dst: &Path, stamp: &Path, ) { let self_contained_dst = &sysroot_dst.join("self-contained"); t!(fs::create_dir_all(&sysroot_dst)); t!(fs::create_dir_all(&sysroot_host_dst)); t!(fs::create_dir_all(&self_contained_dst)); for (path, dependency_type) in builder.read_stamp_file(stamp) { let dst = match dependency_type { DependencyType::Host => sysroot_host_dst, DependencyType::Target => sysroot_dst, DependencyType::TargetSelfContained => self_contained_dst, }; builder.copy(&path, &dst.join(path.file_name().unwrap())); } } pub fn run_cargo( builder: &Builder<'_>, cargo: Cargo, tail_args: Vec, stamp: &Path, additional_target_deps: Vec<(PathBuf, DependencyType)>, is_check: bool, rlib_only_metadata: bool, ) -> Vec { if builder.config.dry_run() { return Vec::new(); } // `target_root_dir` looks like $dir/$target/release let target_root_dir = stamp.parent().unwrap(); // `target_deps_dir` looks like $dir/$target/release/deps let target_deps_dir = target_root_dir.join("deps"); // `host_root_dir` looks like $dir/release let host_root_dir = target_root_dir .parent() .unwrap() // chop off `release` .parent() .unwrap() // chop off `$target` .join(target_root_dir.file_name().unwrap()); // Spawn Cargo slurping up its JSON output. We'll start building up the // `deps` array of all files it generated along with a `toplevel` array of // files we need to probe for later. let mut deps = Vec::new(); let mut toplevel = Vec::new(); let ok = stream_cargo(builder, cargo, tail_args, &mut |msg| { let (filenames, crate_types) = match msg { CargoMessage::CompilerArtifact { filenames, target: CargoTarget { crate_types }, .. } => (filenames, crate_types), _ => return, }; for filename in filenames { // Skip files like executables let mut keep = false; if filename.ends_with(".lib") || filename.ends_with(".a") || is_debug_info(&filename) || is_dylib(&filename) { // Always keep native libraries, rust dylibs and debuginfo keep = true; } if is_check && filename.ends_with(".rmeta") { // During check builds we need to keep crate metadata keep = true; } else if rlib_only_metadata { if filename.contains("jemalloc_sys") || filename.contains("rustc_smir") { // jemalloc_sys and rustc_smir are not linked into librustc_driver.so, // so we need to distribute them as rlib to be able to use them. keep |= filename.ends_with(".rlib"); } else { // Distribute the rest of the rustc crates as rmeta files only to reduce // the tarball sizes by about 50%. The object files are linked into // librustc_driver.so, so it is still possible to link against them. keep |= filename.ends_with(".rmeta"); } } else { // In all other cases keep all rlibs keep |= filename.ends_with(".rlib"); } if !keep { continue; } let filename = Path::new(&*filename); // If this was an output file in the "host dir" we don't actually // worry about it, it's not relevant for us if filename.starts_with(&host_root_dir) { // Unless it's a proc macro used in the compiler if crate_types.iter().any(|t| t == "proc-macro") { deps.push((filename.to_path_buf(), DependencyType::Host)); } continue; } // If this was output in the `deps` dir then this is a precise file // name (hash included) so we start tracking it. if filename.starts_with(&target_deps_dir) { deps.push((filename.to_path_buf(), DependencyType::Target)); continue; } // Otherwise this was a "top level artifact" which right now doesn't // have a hash in the name, but there's a version of this file in // the `deps` folder which *does* have a hash in the name. That's // the one we'll want to we'll probe for it later. // // We do not use `Path::file_stem` or `Path::extension` here, // because some generated files may have multiple extensions e.g. // `std-.dll.lib` on Windows. The aforementioned methods only // split the file name by the last extension (`.lib`) while we need // to split by all extensions (`.dll.lib`). let expected_len = t!(filename.metadata()).len(); let filename = filename.file_name().unwrap().to_str().unwrap(); let mut parts = filename.splitn(2, '.'); let file_stem = parts.next().unwrap().to_owned(); let extension = parts.next().unwrap().to_owned(); toplevel.push((file_stem, extension, expected_len)); } }); if !ok { crate::exit!(1); } // Ok now we need to actually find all the files listed in `toplevel`. We've // got a list of prefix/extensions and we basically just need to find the // most recent file in the `deps` folder corresponding to each one. let contents = t!(target_deps_dir.read_dir()) .map(|e| t!(e)) .map(|e| (e.path(), e.file_name().into_string().unwrap(), t!(e.metadata()))) .collect::>(); for (prefix, extension, expected_len) in toplevel { let candidates = contents.iter().filter(|&&(_, ref filename, ref meta)| { meta.len() == expected_len && filename .strip_prefix(&prefix[..]) .map(|s| s.starts_with('-') && s.ends_with(&extension[..])) .unwrap_or(false) }); let max = candidates.max_by_key(|&&(_, _, ref metadata)| { metadata.modified().expect("mtime should be available on all relevant OSes") }); let path_to_add = match max { Some(triple) => triple.0.to_str().unwrap(), None => panic!("no output generated for {prefix:?} {extension:?}"), }; if is_dylib(path_to_add) { let candidate = format!("{path_to_add}.lib"); let candidate = PathBuf::from(candidate); if candidate.exists() { deps.push((candidate, DependencyType::Target)); } } deps.push((path_to_add.into(), DependencyType::Target)); } deps.extend(additional_target_deps); deps.sort(); let mut new_contents = Vec::new(); for (dep, dependency_type) in deps.iter() { new_contents.extend(match *dependency_type { DependencyType::Host => b"h", DependencyType::Target => b"t", DependencyType::TargetSelfContained => b"s", }); new_contents.extend(dep.to_str().unwrap().as_bytes()); new_contents.extend(b"\0"); } t!(fs::write(&stamp, &new_contents)); deps.into_iter().map(|(d, _)| d).collect() } pub fn stream_cargo( builder: &Builder<'_>, cargo: Cargo, tail_args: Vec, cb: &mut dyn FnMut(CargoMessage<'_>), ) -> bool { let mut cargo = Command::from(cargo); if builder.config.dry_run() { return true; } // Instruct Cargo to give us json messages on stdout, critically leaving // stderr as piped so we can get those pretty colors. let mut message_format = if builder.config.json_output { String::from("json") } else { String::from("json-render-diagnostics") }; if let Some(s) = &builder.config.rustc_error_format { message_format.push_str(",json-diagnostic-"); message_format.push_str(s); } cargo.arg("--message-format").arg(message_format).stdout(Stdio::piped()); for arg in tail_args { cargo.arg(arg); } builder.verbose(&format!("running: {cargo:?}")); let mut child = match cargo.spawn() { Ok(child) => child, Err(e) => panic!("failed to execute command: {cargo:?}\nerror: {e}"), }; // Spawn Cargo slurping up its JSON output. We'll start building up the // `deps` array of all files it generated along with a `toplevel` array of // files we need to probe for later. let stdout = BufReader::new(child.stdout.take().unwrap()); for line in stdout.lines() { let line = t!(line); match serde_json::from_str::>(&line) { Ok(msg) => { if builder.config.json_output { // Forward JSON to stdout. println!("{line}"); } cb(msg) } // If this was informational, just print it out and continue Err(_) => println!("{line}"), } } // Make sure Cargo actually succeeded after we read all of its stdout. let status = t!(child.wait()); if builder.is_verbose() && !status.success() { eprintln!( "command did not execute successfully: {cargo:?}\n\ expected success, got: {status}" ); } status.success() } #[derive(Deserialize)] pub struct CargoTarget<'a> { crate_types: Vec>, } #[derive(Deserialize)] #[serde(tag = "reason", rename_all = "kebab-case")] pub enum CargoMessage<'a> { CompilerArtifact { package_id: Cow<'a, str>, features: Vec>, filenames: Vec>, target: CargoTarget<'a>, }, BuildScriptExecuted { package_id: Cow<'a, str>, }, BuildFinished { success: bool, }, } pub fn strip_debug(builder: &Builder<'_>, target: TargetSelection, path: &Path) { // FIXME: to make things simpler for now, limit this to the host and target where we know // `strip -g` is both available and will fix the issue, i.e. on a x64 linux host that is not // cross-compiling. Expand this to other appropriate targets in the future. if target != "x86_64-unknown-linux-gnu" || target != builder.config.build || !path.exists() { return; } let previous_mtime = FileTime::from_last_modification_time(&path.metadata().unwrap()); // Note: `output` will propagate any errors here. output(Command::new("strip").arg("--strip-debug").arg(path)); // After running `strip`, we have to set the file modification time to what it was before, // otherwise we risk Cargo invalidating its fingerprint and rebuilding the world next time // bootstrap is invoked. // // An example of this is if we run this on librustc_driver.so. In the first invocation: // - Cargo will build librustc_driver.so (mtime of 1) // - Cargo will build rustc-main (mtime of 2) // - Bootstrap will strip librustc_driver.so (changing the mtime to 3). // // In the second invocation of bootstrap, Cargo will see that the mtime of librustc_driver.so // is greater than the mtime of rustc-main, and will rebuild rustc-main. That will then cause // everything else (standard library, future stages...) to be rebuilt. t!(filetime::set_file_mtime(path, previous_mtime)); }