//! Logic to invoke `cargo` for building build-dependencies (build scripts and proc-macros) as well as //! executing the build scripts to fetch required dependency information (`OUT_DIR` env var, extra //! cfg flags, etc). //! //! In essence this just invokes `cargo` with the appropriate output format which we consume, //! but if enabled we will also use `RUSTC_WRAPPER` to only compile the build scripts and //! proc-macros and skip everything else. use std::{cell::RefCell, io, mem, process::Command}; use base_db::Env; use cargo_metadata::{camino::Utf8Path, Message}; use cfg::CfgAtom; use itertools::Itertools; use la_arena::ArenaMap; use paths::{AbsPath, AbsPathBuf, Utf8PathBuf}; use rustc_hash::{FxHashMap, FxHashSet}; use serde::Deserialize; use toolchain::Tool; use crate::{ utf8_stdout, CargoConfig, CargoFeatures, CargoWorkspace, InvocationStrategy, ManifestPath, Package, Sysroot, TargetKind, }; /// Output of the build script and proc-macro building steps for a workspace. #[derive(Debug, Default, Clone, PartialEq, Eq)] pub struct WorkspaceBuildScripts { outputs: ArenaMap, error: Option, } /// Output of the build script and proc-macro building step for a concrete package. #[derive(Debug, Clone, Default, PartialEq, Eq)] pub(crate) struct BuildScriptOutput { /// List of config flags defined by this package's build script. pub(crate) cfgs: Vec, /// List of cargo-related environment variables with their value. /// /// If the package has a build script which defines environment variables, /// they can also be found here. pub(crate) envs: Env, /// Directory where a build script might place its output. pub(crate) out_dir: Option, /// Path to the proc-macro library file if this package exposes proc-macros. pub(crate) proc_macro_dylib_path: Option, } impl BuildScriptOutput { fn is_empty(&self) -> bool { self.cfgs.is_empty() && self.envs.is_empty() && self.out_dir.is_none() && self.proc_macro_dylib_path.is_none() } } impl WorkspaceBuildScripts { /// Runs the build scripts for the given workspace pub(crate) fn run_for_workspace( config: &CargoConfig, workspace: &CargoWorkspace, progress: &dyn Fn(String), sysroot: &Sysroot, ) -> io::Result { let current_dir = workspace.workspace_root(); let allowed_features = workspace.workspace_features(); let cmd = Self::build_command( config, &allowed_features, workspace.manifest_path(), current_dir, sysroot, )?; Self::run_per_ws(cmd, workspace, progress) } /// Runs the build scripts by invoking the configured command *once*. /// This populates the outputs for all passed in workspaces. pub(crate) fn run_once( config: &CargoConfig, workspaces: &[&CargoWorkspace], progress: &dyn Fn(String), working_directory: &AbsPathBuf, ) -> io::Result> { assert_eq!(config.invocation_strategy, InvocationStrategy::Once); let cmd = Self::build_command( config, &Default::default(), // This is not gonna be used anyways, so just construct a dummy here &ManifestPath::try_from(working_directory.clone()).unwrap(), working_directory, &Sysroot::empty(), )?; // NB: Cargo.toml could have been modified between `cargo metadata` and // `cargo check`. We shouldn't assume that package ids we see here are // exactly those from `config`. let mut by_id = FxHashMap::default(); // some workspaces might depend on the same crates, so we need to duplicate the outputs // to those collisions let mut collisions = Vec::new(); let mut res: Vec<_> = workspaces .iter() .enumerate() .map(|(idx, workspace)| { let mut res = WorkspaceBuildScripts::default(); for package in workspace.packages() { res.outputs.insert(package, BuildScriptOutput::default()); if by_id.contains_key(&workspace[package].id) { collisions.push((&workspace[package].id, idx, package)); } else { by_id.insert(workspace[package].id.clone(), (package, idx)); } } res }) .collect(); let errors = Self::run_command( cmd, |package, cb| { if let Some(&(package, workspace)) = by_id.get(package) { cb(&workspaces[workspace][package].name, &mut res[workspace].outputs[package]); } }, progress, )?; res.iter_mut().for_each(|it| it.error.clone_from(&errors)); collisions.into_iter().for_each(|(id, workspace, package)| { if let Some(&(p, w)) = by_id.get(id) { res[workspace].outputs[package] = res[w].outputs[p].clone(); } }); if tracing::enabled!(tracing::Level::INFO) { for (idx, workspace) in workspaces.iter().enumerate() { for package in workspace.packages() { let package_build_data = &mut res[idx].outputs[package]; if !package_build_data.is_empty() { tracing::info!( "{}: {package_build_data:?}", workspace[package].manifest.parent(), ); } } } } Ok(res) } pub fn error(&self) -> Option<&str> { self.error.as_deref() } pub(crate) fn get_output(&self, idx: Package) -> Option<&BuildScriptOutput> { self.outputs.get(idx) } /// Assembles build script outputs for the rustc crates via `--print target-libdir`. pub(crate) fn rustc_crates( rustc: &CargoWorkspace, current_dir: &AbsPath, extra_env: &FxHashMap, sysroot: &Sysroot, ) -> Self { let mut bs = WorkspaceBuildScripts::default(); for p in rustc.packages() { bs.outputs.insert(p, BuildScriptOutput::default()); } let res = (|| { let target_libdir = (|| { let mut cargo_config = sysroot.tool(Tool::Cargo); cargo_config.envs(extra_env); cargo_config .current_dir(current_dir) .args(["rustc", "-Z", "unstable-options", "--print", "target-libdir"]) .env("RUSTC_BOOTSTRAP", "1"); if let Ok(it) = utf8_stdout(cargo_config) { return Ok(it); } let mut cmd = sysroot.tool(Tool::Rustc); cmd.envs(extra_env); cmd.args(["--print", "target-libdir"]); utf8_stdout(cmd) })()?; let target_libdir = AbsPathBuf::try_from(Utf8PathBuf::from(target_libdir)) .map_err(|_| anyhow::format_err!("target-libdir was not an absolute path"))?; tracing::info!("Loading rustc proc-macro paths from {target_libdir}"); let proc_macro_dylibs: Vec<(String, AbsPathBuf)> = std::fs::read_dir(target_libdir)? .filter_map(|entry| { let dir_entry = entry.ok()?; if dir_entry.file_type().ok()?.is_file() { let path = dir_entry.path(); let extension = path.extension()?; if extension == std::env::consts::DLL_EXTENSION { let name = path.file_stem()?.to_str()?.split_once('-')?.0.to_owned(); let path = AbsPathBuf::try_from(Utf8PathBuf::from_path_buf(path).ok()?) .ok()?; return Some((name, path)); } } None }) .collect(); for p in rustc.packages() { let package = &rustc[p]; if package .targets .iter() .any(|&it| matches!(rustc[it].kind, TargetKind::Lib { is_proc_macro: true })) { if let Some((_, path)) = proc_macro_dylibs .iter() .find(|(name, _)| *name.trim_start_matches("lib") == package.name) { bs.outputs[p].proc_macro_dylib_path = Some(path.clone()); } } } if tracing::enabled!(tracing::Level::INFO) { for package in rustc.packages() { let package_build_data = &bs.outputs[package]; if !package_build_data.is_empty() { tracing::info!( "{}: {package_build_data:?}", rustc[package].manifest.parent(), ); } } } Ok(()) })(); if let Err::<_, anyhow::Error>(e) = res { bs.error = Some(e.to_string()); } bs } fn run_per_ws( cmd: Command, workspace: &CargoWorkspace, progress: &dyn Fn(String), ) -> io::Result { let mut res = WorkspaceBuildScripts::default(); let outputs = &mut res.outputs; // NB: Cargo.toml could have been modified between `cargo metadata` and // `cargo check`. We shouldn't assume that package ids we see here are // exactly those from `config`. let mut by_id: FxHashMap = FxHashMap::default(); for package in workspace.packages() { outputs.insert(package, BuildScriptOutput::default()); by_id.insert(workspace[package].id.clone(), package); } res.error = Self::run_command( cmd, |package, cb| { if let Some(&package) = by_id.get(package) { cb(&workspace[package].name, &mut outputs[package]); } }, progress, )?; if tracing::enabled!(tracing::Level::INFO) { for package in workspace.packages() { let package_build_data = &outputs[package]; if !package_build_data.is_empty() { tracing::info!( "{}: {package_build_data:?}", workspace[package].manifest.parent(), ); } } } Ok(res) } fn run_command( cmd: Command, // ideally this would be something like: // with_output_for: impl FnMut(&str, dyn FnOnce(&mut BuildScriptOutput)), // but owned trait objects aren't a thing mut with_output_for: impl FnMut(&str, &mut dyn FnMut(&str, &mut BuildScriptOutput)), progress: &dyn Fn(String), ) -> io::Result> { let errors = RefCell::new(String::new()); let push_err = |err: &str| { let mut e = errors.borrow_mut(); e.push_str(err); e.push('\n'); }; tracing::info!("Running build scripts: {:?}", cmd); let output = stdx::process::spawn_with_streaming_output( cmd, &mut |line| { // Copy-pasted from existing cargo_metadata. It seems like we // should be using serde_stacker here? let mut deserializer = serde_json::Deserializer::from_str(line); deserializer.disable_recursion_limit(); let message = Message::deserialize(&mut deserializer) .unwrap_or_else(|_| Message::TextLine(line.to_owned())); match message { Message::BuildScriptExecuted(mut message) => { with_output_for(&message.package_id.repr, &mut |name, data| { progress(format!("running build-script: {name}")); let cfgs = { let mut acc = Vec::new(); for cfg in &message.cfgs { match crate::parse_cfg(cfg) { Ok(it) => acc.push(it), Err(err) => { push_err(&format!( "invalid cfg from cargo-metadata: {err}" )); return; } }; } acc }; data.envs.extend(message.env.drain(..)); // cargo_metadata crate returns default (empty) path for // older cargos, which is not absolute, so work around that. let out_dir = mem::take(&mut message.out_dir); if !out_dir.as_str().is_empty() { let out_dir = AbsPathBuf::assert(out_dir); // inject_cargo_env(package, package_build_data); data.envs.insert("OUT_DIR", out_dir.as_str()); data.out_dir = Some(out_dir); data.cfgs = cfgs; } }); } Message::CompilerArtifact(message) => { with_output_for(&message.package_id.repr, &mut |name, data| { progress(format!("building proc-macros: {name}")); if message.target.kind.iter().any(|k| k == "proc-macro") { // Skip rmeta file if let Some(filename) = message.filenames.iter().find(|file| is_dylib(file)) { let filename = AbsPath::assert(filename); data.proc_macro_dylib_path = Some(filename.to_owned()); } } }); } Message::CompilerMessage(message) => { progress(message.target.name); if let Some(diag) = message.message.rendered.as_deref() { push_err(diag); } } Message::BuildFinished(_) => {} Message::TextLine(_) => {} _ => {} } }, &mut |line| { push_err(line); }, )?; let errors = if !output.status.success() { let errors = errors.into_inner(); Some(if errors.is_empty() { "cargo check failed".to_owned() } else { errors }) } else { None }; Ok(errors) } fn build_command( config: &CargoConfig, allowed_features: &FxHashSet, manifest_path: &ManifestPath, current_dir: &AbsPath, sysroot: &Sysroot, ) -> io::Result { let mut cmd = match config.run_build_script_command.as_deref() { Some([program, args @ ..]) => { let mut cmd = Command::new(program); cmd.args(args); cmd } _ => { let mut cmd = sysroot.tool(Tool::Cargo); cmd.args(["check", "--quiet", "--workspace", "--message-format=json"]); cmd.args(&config.extra_args); cmd.arg("--manifest-path"); cmd.arg(manifest_path); if let Some(target_dir) = &config.target_dir { cmd.arg("--target-dir").arg(target_dir); } // --all-targets includes tests, benches and examples in addition to the // default lib and bins. This is an independent concept from the --target // flag below. if config.all_targets { cmd.arg("--all-targets"); } if let Some(target) = &config.target { cmd.args(["--target", target]); } match &config.features { CargoFeatures::All => { cmd.arg("--all-features"); } CargoFeatures::Selected { features, no_default_features } => { if *no_default_features { cmd.arg("--no-default-features"); } if !features.is_empty() { cmd.arg("--features"); cmd.arg( features .iter() .filter(|&feat| allowed_features.contains(feat)) .join(","), ); } } } if manifest_path.is_rust_manifest() { cmd.arg("-Zscript"); } cmd.arg("--keep-going"); cmd } }; cmd.current_dir(current_dir); cmd.envs(&config.extra_env); if config.wrap_rustc_in_build_scripts { // Setup RUSTC_WRAPPER to point to `rust-analyzer` binary itself. We use // that to compile only proc macros and build scripts during the initial // `cargo check`. let myself = std::env::current_exe()?; cmd.env("RUSTC_WRAPPER", myself); cmd.env("RA_RUSTC_WRAPPER", "1"); } Ok(cmd) } } // FIXME: Find a better way to know if it is a dylib. fn is_dylib(path: &Utf8Path) -> bool { match path.extension().map(|e| e.to_owned().to_lowercase()) { None => false, Some(ext) => matches!(ext.as_str(), "dll" | "dylib" | "so"), } }