Skip to main content

cargo/core/compiler/
mod.rs

1//! # Interact with the compiler
2//!
3//! If you consider [`ops::cargo_compile::compile`] as a `rustc` driver but on
4//! Cargo side, this module is kinda the `rustc_interface` for that merits.
5//! It contains all the interaction between Cargo and the rustc compiler,
6//! from preparing the context for the entire build process, to scheduling
7//! and executing each unit of work (e.g. running `rustc`), to managing and
8//! caching the output artifact of a build.
9//!
10//! However, it hasn't yet exposed a clear definition of each phase or session,
11//! like what rustc has done. Also, no one knows if Cargo really needs that.
12//! To be pragmatic, here we list a handful of items you may want to learn:
13//!
14//! * [`BuildContext`] is a static context containing all information you need
15//!   before a build gets started.
16//! * [`BuildRunner`] is the center of the world, coordinating a running build and
17//!   collecting information from it.
18//! * [`custom_build`] is the home of build script executions and output parsing.
19//! * [`fingerprint`] not only defines but also executes a set of rules to
20//!   determine if a re-compile is needed.
21//! * [`job_queue`] is where the parallelism, job scheduling, and communication
22//!   machinery happen between Cargo and the compiler.
23//! * [`layout`] defines and manages output artifacts of a build in the filesystem.
24//! * [`unit_dependencies`] is for building a dependency graph for compilation
25//!   from a result of dependency resolution.
26//! * [`Unit`] contains sufficient information to build something, usually
27//!   turning into a compiler invocation in a later phase.
28//!
29//! [`ops::cargo_compile::compile`]: crate::ops::compile
30
31pub mod artifact;
32mod build_config;
33pub(crate) mod build_context;
34pub(crate) mod build_runner;
35mod compilation;
36mod compile_kind;
37mod crate_type;
38mod custom_build;
39pub(crate) mod fingerprint;
40pub mod future_incompat;
41pub(crate) mod job_queue;
42pub(crate) mod layout;
43mod links;
44mod locking;
45mod lto;
46mod output_depinfo;
47mod output_sbom;
48pub mod rustdoc;
49pub mod standard_lib;
50pub mod timings;
51mod unit;
52pub mod unit_dependencies;
53pub mod unit_graph;
54
55use std::borrow::Cow;
56use std::cell::OnceCell;
57use std::collections::{BTreeMap, HashMap, HashSet};
58use std::env;
59use std::ffi::{OsStr, OsString};
60use std::fmt::Display;
61use std::fs::{self, File};
62use std::io::{BufRead, BufWriter, Write};
63use std::ops::Range;
64use std::path::{Path, PathBuf};
65use std::sync::{Arc, LazyLock};
66
67use annotate_snippets::{AnnotationKind, Group, Level, Renderer, Snippet};
68use anyhow::{Context as _, Error};
69use cargo_platform::{Cfg, Platform};
70use itertools::Itertools;
71use regex::Regex;
72use tracing::{debug, instrument, trace};
73
74pub use self::build_config::UserIntent;
75pub use self::build_config::{BuildConfig, CompileMode, MessageFormat};
76pub use self::build_context::BuildContext;
77pub use self::build_context::FileFlavor;
78pub use self::build_context::FileType;
79pub use self::build_context::RustcTargetData;
80pub use self::build_context::TargetInfo;
81pub use self::build_runner::{BuildRunner, Metadata, UnitHash};
82pub use self::compilation::{Compilation, Doctest, UnitOutput};
83pub use self::compile_kind::{CompileKind, CompileKindFallback, CompileTarget};
84pub use self::crate_type::CrateType;
85pub use self::custom_build::LinkArgTarget;
86pub use self::custom_build::{BuildOutput, BuildScriptOutputs, BuildScripts, LibraryPath};
87pub(crate) use self::fingerprint::DirtyReason;
88pub use self::fingerprint::RustdocFingerprint;
89pub use self::job_queue::Freshness;
90use self::job_queue::{Job, JobQueue, JobState, Work};
91pub(crate) use self::layout::Layout;
92pub use self::lto::Lto;
93use self::output_depinfo::output_depinfo;
94use self::output_sbom::build_sbom;
95use self::unit_graph::UnitDep;
96
97use crate::core::compiler::future_incompat::FutureIncompatReport;
98use crate::core::compiler::locking::LockKey;
99use crate::core::compiler::timings::SectionTiming;
100pub use crate::core::compiler::unit::Unit;
101pub use crate::core::compiler::unit::UnitIndex;
102pub use crate::core::compiler::unit::UnitInterner;
103use crate::core::manifest::TargetSourcePath;
104use crate::core::profiles::{PanicStrategy, Profile, StripInner};
105use crate::core::{Feature, PackageId, Target, Verbosity};
106use crate::lints::get_key_value;
107use crate::util::OnceExt;
108use crate::util::context::WarningHandling;
109use crate::util::errors::{CargoResult, VerboseError};
110use crate::util::interning::InternedString;
111use crate::util::machine_message::{self, Message};
112use crate::util::{add_path_args, internal, path_args};
113
114use cargo_util::{ProcessBuilder, ProcessError, paths};
115use cargo_util_schemas::manifest::TomlDebugInfo;
116use cargo_util_schemas::manifest::TomlTrimPaths;
117use cargo_util_schemas::manifest::TomlTrimPathsValue;
118use rustfix::diagnostics::Applicability;
119
120const RUSTDOC_CRATE_VERSION_FLAG: &str = "--crate-version";
121
122/// A glorified callback for executing calls to rustc. Rather than calling rustc
123/// directly, we'll use an `Executor`, giving clients an opportunity to intercept
124/// the build calls.
125pub trait Executor: Send + Sync + 'static {
126    /// Called after a rustc process invocation is prepared up-front for a given
127    /// unit of work (may still be modified for runtime-known dependencies, when
128    /// the work is actually executed).
129    fn init(&self, _build_runner: &BuildRunner<'_, '_>, _unit: &Unit) {}
130
131    /// In case of an `Err`, Cargo will not continue with the build process for
132    /// this package.
133    fn exec(
134        &self,
135        cmd: &ProcessBuilder,
136        id: PackageId,
137        target: &Target,
138        mode: CompileMode,
139        on_stdout_line: &mut dyn FnMut(&str) -> CargoResult<()>,
140        on_stderr_line: &mut dyn FnMut(&str) -> CargoResult<()>,
141    ) -> CargoResult<()>;
142
143    /// Queried when queuing each unit of work. If it returns true, then the
144    /// unit will always be rebuilt, independent of whether it needs to be.
145    fn force_rebuild(&self, _unit: &Unit) -> bool {
146        false
147    }
148}
149
150/// A `DefaultExecutor` calls rustc without doing anything else. It is Cargo's
151/// default behaviour.
152#[derive(Copy, Clone)]
153pub struct DefaultExecutor;
154
155impl Executor for DefaultExecutor {
156    #[instrument(name = "rustc", skip_all, fields(package = id.name().as_str(), process = cmd.to_string()))]
157    fn exec(
158        &self,
159        cmd: &ProcessBuilder,
160        id: PackageId,
161        _target: &Target,
162        _mode: CompileMode,
163        on_stdout_line: &mut dyn FnMut(&str) -> CargoResult<()>,
164        on_stderr_line: &mut dyn FnMut(&str) -> CargoResult<()>,
165    ) -> CargoResult<()> {
166        cmd.exec_with_streaming(on_stdout_line, on_stderr_line, false)
167            .map(drop)
168    }
169}
170
171/// Builds up and enqueue a list of pending jobs onto the `job` queue.
172///
173/// Starting from the `unit`, this function recursively calls itself to build
174/// all jobs for dependencies of the `unit`. Each of these jobs represents
175/// compiling a particular package.
176///
177/// Note that **no actual work is executed as part of this**, that's all done
178/// next as part of [`JobQueue::execute`] function which will run everything
179/// in order with proper parallelism.
180#[tracing::instrument(skip(build_runner, jobs, exec))]
181fn compile<'gctx>(
182    build_runner: &mut BuildRunner<'_, 'gctx>,
183    jobs: &mut JobQueue<'gctx>,
184    unit: &Unit,
185    exec: &Arc<dyn Executor>,
186    force_rebuild: bool,
187) -> CargoResult<()> {
188    let bcx = build_runner.bcx;
189    if !build_runner.compiled.insert(unit.clone()) {
190        return Ok(());
191    }
192
193    let lock = if build_runner.bcx.gctx.cli_unstable().fine_grain_locking {
194        Some(build_runner.lock_manager.lock_shared(build_runner, unit)?)
195    } else {
196        None
197    };
198
199    // If we are in `--compile-time-deps` and the given unit is not a compile time
200    // dependency, skip compiling the unit and jumps to dependencies, which still
201    // have chances to be compile time dependencies
202    if !unit.skip_non_compile_time_dep {
203        // Build up the work to be done to compile this unit, enqueuing it once
204        // we've got everything constructed.
205        fingerprint::prepare_init(build_runner, unit)?;
206
207        let job = if unit.mode.is_run_custom_build() {
208            custom_build::prepare(build_runner, unit)?
209        } else if unit.mode.is_doc_test() {
210            // We run these targets later, so this is just a no-op for now.
211            Job::new_fresh()
212        } else {
213            let force = exec.force_rebuild(unit) || force_rebuild;
214            let mut job = fingerprint::prepare_target(build_runner, unit, force)?;
215            job.before(if job.freshness().is_dirty() {
216                let work = if unit.mode.is_doc() || unit.mode.is_doc_scrape() {
217                    rustdoc(build_runner, unit)?
218                } else {
219                    rustc(build_runner, unit, exec)?
220                };
221                work.then(link_targets(build_runner, unit, false)?)
222            } else {
223                // We always replay the output cache,
224                // since it might contain future-incompat-report messages
225                let show_diagnostics = unit.show_warnings(bcx.gctx)
226                    && build_runner.bcx.gctx.warning_handling()? != WarningHandling::Allow;
227                let manifest = ManifestErrorContext::new(build_runner, unit);
228                let work = replay_output_cache(
229                    unit.pkg.package_id(),
230                    manifest,
231                    &unit.target,
232                    build_runner.files().message_cache_path(unit),
233                    build_runner.bcx.build_config.message_format,
234                    show_diagnostics,
235                );
236                // Need to link targets on both the dirty and fresh.
237                work.then(link_targets(build_runner, unit, true)?)
238            });
239
240            // If -Zfine-grain-locking is enabled, we wrap the job with an upgrade to exclusive
241            // lock before starting, then downgrade to a shared lock after the job is finished.
242            if build_runner.bcx.gctx.cli_unstable().fine_grain_locking && job.freshness().is_dirty()
243            {
244                if let Some(lock) = lock {
245                    // Here we unlock the current shared lock to avoid deadlocking with other cargo
246                    // processes. Then we configure our compile job to take an exclusive lock
247                    // before starting. Once we are done compiling (including both rmeta and rlib)
248                    // we downgrade to a shared lock to allow other cargo's to read the build unit.
249                    // We will hold this shared lock for the remainder of compilation to prevent
250                    // other cargo from re-compiling while we are still using the unit.
251                    build_runner.lock_manager.unlock(&lock)?;
252                    job.before(prebuild_lock_exclusive(lock.clone()));
253                    job.after(downgrade_lock_to_shared(lock));
254                }
255            }
256
257            job
258        };
259        jobs.enqueue(build_runner, unit, job)?;
260    }
261
262    // Be sure to compile all dependencies of this target as well.
263    let deps = Vec::from(build_runner.unit_deps(unit)); // Create vec due to mutable borrow.
264    for dep in deps {
265        compile(build_runner, jobs, &dep.unit, exec, false)?;
266    }
267
268    Ok(())
269}
270
271/// Generates the warning message used when fallible doc-scrape units fail,
272/// either for rustdoc or rustc.
273fn make_failed_scrape_diagnostic(
274    build_runner: &BuildRunner<'_, '_>,
275    unit: &Unit,
276    top_line: impl Display,
277) -> String {
278    let manifest_path = unit.pkg.manifest_path();
279    let relative_manifest_path = manifest_path
280        .strip_prefix(build_runner.bcx.ws.root())
281        .unwrap_or(&manifest_path);
282
283    format!(
284        "\
285{top_line}
286    Try running with `--verbose` to see the error message.
287    If an example should not be scanned, then consider adding `doc-scrape-examples = false` to its `[[example]]` definition in {}",
288        relative_manifest_path.display()
289    )
290}
291
292/// Creates a unit of work invoking `rustc` for building the `unit`.
293fn rustc(
294    build_runner: &mut BuildRunner<'_, '_>,
295    unit: &Unit,
296    exec: &Arc<dyn Executor>,
297) -> CargoResult<Work> {
298    let mut rustc = prepare_rustc(build_runner, unit)?;
299
300    let name = unit.pkg.name();
301
302    let outputs = build_runner.outputs(unit)?;
303    let root = build_runner.files().output_dir(unit);
304
305    // Prepare the native lib state (extra `-L` and `-l` flags).
306    let build_script_outputs = Arc::clone(&build_runner.build_script_outputs);
307    let current_id = unit.pkg.package_id();
308    let manifest = ManifestErrorContext::new(build_runner, unit);
309    let build_scripts = build_runner.build_scripts.get(unit).cloned();
310
311    // If we are a binary and the package also contains a library, then we
312    // don't pass the `-l` flags.
313    let pass_l_flag = unit.target.is_lib() || !unit.pkg.targets().iter().any(|t| t.is_lib());
314
315    let dep_info_name =
316        if let Some(c_extra_filename) = build_runner.files().metadata(unit).c_extra_filename() {
317            format!("{}-{}.d", unit.target.crate_name(), c_extra_filename)
318        } else {
319            format!("{}.d", unit.target.crate_name())
320        };
321    let rustc_dep_info_loc = root.join(dep_info_name);
322    let dep_info_loc = fingerprint::dep_info_loc(build_runner, unit);
323
324    let mut output_options = OutputOptions::new(build_runner, unit);
325    let package_id = unit.pkg.package_id();
326    let target = Target::clone(&unit.target);
327    let mode = unit.mode;
328
329    exec.init(build_runner, unit);
330    let exec = exec.clone();
331
332    let root_output = build_runner.files().host_dest().map(|v| v.to_path_buf());
333    let build_dir = build_runner.bcx.ws.build_dir().into_path_unlocked();
334    let pkg_root = unit.pkg.root().to_path_buf();
335    let cwd = rustc
336        .get_cwd()
337        .unwrap_or_else(|| build_runner.bcx.gctx.cwd())
338        .to_path_buf();
339    let fingerprint_dir = build_runner.files().fingerprint_dir(unit);
340    let script_metadatas = build_runner.find_build_script_metadatas(unit);
341    let is_local = unit.is_local();
342    let artifact = unit.artifact;
343    let sbom_files = build_runner.sbom_output_files(unit)?;
344    let sbom = build_sbom(build_runner, unit)?;
345
346    let hide_diagnostics_for_scrape_unit = build_runner.bcx.unit_can_fail_for_docscraping(unit)
347        && !matches!(
348            build_runner.bcx.gctx.shell().verbosity(),
349            Verbosity::Verbose
350        );
351    let failed_scrape_diagnostic = hide_diagnostics_for_scrape_unit.then(|| {
352        // If this unit is needed for doc-scraping, then we generate a diagnostic that
353        // describes the set of reverse-dependencies that cause the unit to be needed.
354        let target_desc = unit.target.description_named();
355        let mut for_scrape_units = build_runner
356            .bcx
357            .scrape_units_have_dep_on(unit)
358            .into_iter()
359            .map(|unit| unit.target.description_named())
360            .collect::<Vec<_>>();
361        for_scrape_units.sort();
362        let for_scrape_units = for_scrape_units.join(", ");
363        make_failed_scrape_diagnostic(build_runner, unit, format_args!("failed to check {target_desc} in package `{name}` as a prerequisite for scraping examples from: {for_scrape_units}"))
364    });
365    if hide_diagnostics_for_scrape_unit {
366        output_options.show_diagnostics = false;
367    }
368    let env_config = Arc::clone(build_runner.bcx.gctx.env_config()?);
369    return Ok(Work::new(move |state| {
370        // Artifacts are in a different location than typical units,
371        // hence we must assure the crate- and target-dependent
372        // directory is present.
373        if artifact.is_true() {
374            paths::create_dir_all(&root)?;
375        }
376
377        // Only at runtime have we discovered what the extra -L and -l
378        // arguments are for native libraries, so we process those here. We
379        // also need to be sure to add any -L paths for our plugins to the
380        // dynamic library load path as a plugin's dynamic library may be
381        // located somewhere in there.
382        // Finally, if custom environment variables have been produced by
383        // previous build scripts, we include them in the rustc invocation.
384        if let Some(build_scripts) = build_scripts {
385            let script_outputs = build_script_outputs.lock().unwrap();
386            add_native_deps(
387                &mut rustc,
388                &script_outputs,
389                &build_scripts,
390                pass_l_flag,
391                &target,
392                current_id,
393                mode,
394            )?;
395            if let Some(ref root_output) = root_output {
396                add_plugin_deps(&mut rustc, &script_outputs, &build_scripts, root_output)?;
397            }
398            add_custom_flags(&mut rustc, &script_outputs, script_metadatas)?;
399        }
400
401        for output in outputs.iter() {
402            // If there is both an rmeta and rlib, rustc will prefer to use the
403            // rlib, even if it is older. Therefore, we must delete the rlib to
404            // force using the new rmeta.
405            if output.path.extension() == Some(OsStr::new("rmeta")) {
406                let dst = root.join(&output.path).with_extension("rlib");
407                if dst.exists() {
408                    paths::remove_file(&dst)?;
409                }
410            }
411
412            // Some linkers do not remove the executable, but truncate and modify it.
413            // That results in the old hard-link being modified even after renamed.
414            // We delete the old artifact here to prevent this behavior from confusing users.
415            // See rust-lang/cargo#8348.
416            if output.hardlink.is_some() && output.path.exists() {
417                _ = paths::remove_file(&output.path).map_err(|e| {
418                    tracing::debug!(
419                        "failed to delete previous output file `{:?}`: {e:?}",
420                        output.path
421                    );
422                });
423            }
424        }
425
426        state.running(&rustc);
427        let timestamp = paths::set_invocation_time(&fingerprint_dir)?;
428        for file in sbom_files {
429            tracing::debug!("writing sbom to {}", file.display());
430            let outfile = BufWriter::new(paths::create(&file)?);
431            serde_json::to_writer(outfile, &sbom)?;
432        }
433
434        let result = exec
435            .exec(
436                &rustc,
437                package_id,
438                &target,
439                mode,
440                &mut |line| on_stdout_line(state, line, package_id, &target),
441                &mut |line| {
442                    on_stderr_line(
443                        state,
444                        line,
445                        package_id,
446                        &manifest,
447                        &target,
448                        &mut output_options,
449                    )
450                },
451            )
452            .map_err(|e| {
453                if output_options.errors_seen == 0 {
454                    // If we didn't expect an error, do not require --verbose to fail.
455                    // This is intended to debug
456                    // https://github.com/rust-lang/crater/issues/733, where we are seeing
457                    // Cargo exit unsuccessfully while seeming to not show any errors.
458                    e
459                } else {
460                    verbose_if_simple_exit_code(e)
461                }
462            })
463            .with_context(|| {
464                // adapted from rustc_errors/src/lib.rs
465                let warnings = match output_options.warnings_seen {
466                    0 => String::new(),
467                    1 => "; 1 warning emitted".to_string(),
468                    count => format!("; {} warnings emitted", count),
469                };
470                let errors = match output_options.errors_seen {
471                    0 => String::new(),
472                    1 => " due to 1 previous error".to_string(),
473                    count => format!(" due to {} previous errors", count),
474                };
475                let name = descriptive_pkg_name(&name, &target, &mode);
476                format!("could not compile {name}{errors}{warnings}")
477            });
478
479        if let Err(e) = result {
480            if let Some(diagnostic) = failed_scrape_diagnostic {
481                state.warning(diagnostic);
482            }
483
484            return Err(e);
485        }
486
487        // Exec should never return with success *and* generate an error.
488        debug_assert_eq!(output_options.errors_seen, 0);
489
490        if rustc_dep_info_loc.exists() {
491            fingerprint::translate_dep_info(
492                &rustc_dep_info_loc,
493                &dep_info_loc,
494                &cwd,
495                &pkg_root,
496                &build_dir,
497                &rustc,
498                // Do not track source files in the fingerprint for registry dependencies.
499                is_local,
500                &env_config,
501            )
502            .with_context(|| {
503                internal(format!(
504                    "could not parse/generate dep info at: {}",
505                    rustc_dep_info_loc.display()
506                ))
507            })?;
508            // This mtime shift allows Cargo to detect if a source file was
509            // modified in the middle of the build.
510            paths::set_file_time_no_err(dep_info_loc, timestamp);
511        }
512
513        // This mtime shift for .rmeta is a workaround as rustc incremental build
514        // since rust-lang/rust#114669 (1.90.0) skips unnecessary rmeta generation.
515        //
516        // The situation is like this:
517        //
518        // 1. When build script execution's external dependendies
519        //    (rerun-if-changed, rerun-if-env-changed) got updated,
520        //    the execution unit reran and got a newer mtime.
521        // 2. rustc type-checked the associated crate, though with incremental
522        //    compilation, no rmeta regeneration. Its `.rmeta` stays old.
523        // 3. Run `cargo check` again. Cargo found build script execution had
524        //    a new mtime than existing crate rmeta, so re-checking the crate.
525        //    However the check is a no-op (input has no change), so stuck.
526        if mode.is_check() {
527            for output in outputs.iter() {
528                paths::set_file_time_no_err(&output.path, timestamp);
529            }
530        }
531
532        Ok(())
533    }));
534
535    // Add all relevant `-L` and `-l` flags from dependencies (now calculated and
536    // present in `state`) to the command provided.
537    fn add_native_deps(
538        rustc: &mut ProcessBuilder,
539        build_script_outputs: &BuildScriptOutputs,
540        build_scripts: &BuildScripts,
541        pass_l_flag: bool,
542        target: &Target,
543        current_id: PackageId,
544        mode: CompileMode,
545    ) -> CargoResult<()> {
546        let mut library_paths = vec![];
547
548        for key in build_scripts.to_link.iter() {
549            let output = build_script_outputs.get(key.1).ok_or_else(|| {
550                internal(format!(
551                    "couldn't find build script output for {}/{}",
552                    key.0, key.1
553                ))
554            })?;
555            library_paths.extend(output.library_paths.iter());
556        }
557
558        // NOTE: This very intentionally does not use the derived ord from LibraryPath because we need to
559        // retain relative ordering within the same type (i.e. not lexicographic). The use of a stable sort
560        // is also important here because it ensures that paths of the same type retain the same relative
561        // ordering (for an unstable sort to work here, the list would need to retain the idx of each element
562        // and then sort by that idx when the type is equivalent.
563        library_paths.sort_by_key(|p| match p {
564            LibraryPath::CargoArtifact(_) => 0,
565            LibraryPath::External(_) => 1,
566        });
567
568        for path in library_paths.iter() {
569            rustc.arg("-L").arg(path.as_ref());
570        }
571
572        for key in build_scripts.to_link.iter() {
573            let output = build_script_outputs.get(key.1).ok_or_else(|| {
574                internal(format!(
575                    "couldn't find build script output for {}/{}",
576                    key.0, key.1
577                ))
578            })?;
579
580            if key.0 == current_id {
581                if pass_l_flag {
582                    for name in output.library_links.iter() {
583                        rustc.arg("-l").arg(name);
584                    }
585                }
586            }
587
588            for (lt, arg) in &output.linker_args {
589                // There was an unintentional change where cdylibs were
590                // allowed to be passed via transitive dependencies. This
591                // clause should have been kept in the `if` block above. For
592                // now, continue allowing it for cdylib only.
593                // See https://github.com/rust-lang/cargo/issues/9562
594                if lt.applies_to(target, mode)
595                    && (key.0 == current_id || *lt == LinkArgTarget::Cdylib)
596                {
597                    rustc.arg("-C").arg(format!("link-arg={}", arg));
598                }
599            }
600        }
601        Ok(())
602    }
603}
604
605fn verbose_if_simple_exit_code(err: Error) -> Error {
606    // If a signal on unix (`code == None`) or an abnormal termination
607    // on Windows (codes like `0xC0000409`), don't hide the error details.
608    match err
609        .downcast_ref::<ProcessError>()
610        .as_ref()
611        .and_then(|perr| perr.code)
612    {
613        Some(n) if cargo_util::is_simple_exit_code(n) => VerboseError::new(err).into(),
614        _ => err,
615    }
616}
617
618fn prebuild_lock_exclusive(lock: LockKey) -> Work {
619    Work::new(move |state| {
620        state.lock_exclusive(&lock)?;
621        Ok(())
622    })
623}
624
625fn downgrade_lock_to_shared(lock: LockKey) -> Work {
626    Work::new(move |state| {
627        state.downgrade_to_shared(&lock)?;
628        Ok(())
629    })
630}
631
632/// Link the compiled target (often of form `foo-{metadata_hash}`) to the
633/// final target. This must happen during both "Fresh" and "Compile".
634fn link_targets(
635    build_runner: &mut BuildRunner<'_, '_>,
636    unit: &Unit,
637    fresh: bool,
638) -> CargoResult<Work> {
639    let bcx = build_runner.bcx;
640    let outputs = build_runner.outputs(unit)?;
641    let export_dir = build_runner.files().export_dir();
642    let package_id = unit.pkg.package_id();
643    let manifest_path = PathBuf::from(unit.pkg.manifest_path());
644    let profile = unit.profile.clone();
645    let unit_mode = unit.mode;
646    let features = unit.features.iter().map(|s| s.to_string()).collect();
647    let json_messages = bcx.build_config.emit_json();
648    let executable = build_runner.get_executable(unit)?;
649    let mut target = Target::clone(&unit.target);
650    if let TargetSourcePath::Metabuild = target.src_path() {
651        // Give it something to serialize.
652        let path = unit
653            .pkg
654            .manifest()
655            .metabuild_path(build_runner.bcx.ws.build_dir());
656        target.set_src_path(TargetSourcePath::Path(path));
657    }
658
659    Ok(Work::new(move |state| {
660        // If we're a "root crate", e.g., the target of this compilation, then we
661        // hard link our outputs out of the `deps` directory into the directory
662        // above. This means that `cargo build` will produce binaries in
663        // `target/debug` which one probably expects.
664        let mut destinations = vec![];
665        for output in outputs.iter() {
666            let src = &output.path;
667            // This may have been a `cargo rustc` command which changes the
668            // output, so the source may not actually exist.
669            if !src.exists() {
670                continue;
671            }
672            let Some(dst) = output.hardlink.as_ref() else {
673                destinations.push(src.clone());
674                continue;
675            };
676            destinations.push(dst.clone());
677            paths::link_or_copy(src, dst)?;
678            if let Some(ref path) = output.export_path {
679                let export_dir = export_dir.as_ref().unwrap();
680                paths::create_dir_all(export_dir)?;
681
682                paths::link_or_copy(src, path)?;
683            }
684        }
685
686        if json_messages {
687            let debuginfo = match profile.debuginfo.into_inner() {
688                TomlDebugInfo::None => machine_message::ArtifactDebuginfo::Int(0),
689                TomlDebugInfo::Limited => machine_message::ArtifactDebuginfo::Int(1),
690                TomlDebugInfo::Full => machine_message::ArtifactDebuginfo::Int(2),
691                TomlDebugInfo::LineDirectivesOnly => {
692                    machine_message::ArtifactDebuginfo::Named("line-directives-only")
693                }
694                TomlDebugInfo::LineTablesOnly => {
695                    machine_message::ArtifactDebuginfo::Named("line-tables-only")
696                }
697            };
698            let art_profile = machine_message::ArtifactProfile {
699                opt_level: profile.opt_level.as_str(),
700                debuginfo: Some(debuginfo),
701                debug_assertions: profile.debug_assertions,
702                overflow_checks: profile.overflow_checks,
703                test: unit_mode.is_any_test(),
704            };
705
706            let msg = machine_message::Artifact {
707                package_id: package_id.to_spec(),
708                manifest_path,
709                target: &target,
710                profile: art_profile,
711                features,
712                filenames: destinations,
713                executable,
714                fresh,
715            }
716            .to_json_string();
717            state.stdout(msg)?;
718        }
719        Ok(())
720    }))
721}
722
723// For all plugin dependencies, add their -L paths (now calculated and present
724// in `build_script_outputs`) to the dynamic library load path for the command
725// to execute.
726fn add_plugin_deps(
727    rustc: &mut ProcessBuilder,
728    build_script_outputs: &BuildScriptOutputs,
729    build_scripts: &BuildScripts,
730    root_output: &Path,
731) -> CargoResult<()> {
732    let var = paths::dylib_path_envvar();
733    let search_path = rustc.get_env(var).unwrap_or_default();
734    let mut search_path = env::split_paths(&search_path).collect::<Vec<_>>();
735    for (pkg_id, metadata) in &build_scripts.plugins {
736        let output = build_script_outputs
737            .get(*metadata)
738            .ok_or_else(|| internal(format!("couldn't find libs for plugin dep {}", pkg_id)))?;
739        search_path.append(&mut filter_dynamic_search_path(
740            output.library_paths.iter().map(AsRef::as_ref),
741            root_output,
742        ));
743    }
744    let search_path = paths::join_paths(&search_path, var)?;
745    rustc.env(var, &search_path);
746    Ok(())
747}
748
749fn get_dynamic_search_path(path: &Path) -> &Path {
750    match path.to_str().and_then(|s| s.split_once("=")) {
751        Some(("native" | "crate" | "dependency" | "framework" | "all", path)) => Path::new(path),
752        _ => path,
753    }
754}
755
756// Determine paths to add to the dynamic search path from -L entries
757//
758// Strip off prefixes like "native=" or "framework=" and filter out directories
759// **not** inside our output directory since they are likely spurious and can cause
760// clashes with system shared libraries (issue #3366).
761fn filter_dynamic_search_path<'a, I>(paths: I, root_output: &Path) -> Vec<PathBuf>
762where
763    I: Iterator<Item = &'a PathBuf>,
764{
765    let mut search_path = vec![];
766    for dir in paths {
767        let dir = get_dynamic_search_path(dir);
768        if dir.starts_with(&root_output) {
769            search_path.push(dir.to_path_buf());
770        } else {
771            debug!(
772                "Not including path {} in runtime library search path because it is \
773                 outside target root {}",
774                dir.display(),
775                root_output.display()
776            );
777        }
778    }
779    search_path
780}
781
782/// Prepares flags and environments we can compute for a `rustc` invocation
783/// before the job queue starts compiling any unit.
784///
785/// This builds a static view of the invocation. Flags depending on the
786/// completion of other units will be added later in runtime, such as flags
787/// from build scripts.
788fn prepare_rustc(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> CargoResult<ProcessBuilder> {
789    let gctx = build_runner.bcx.gctx;
790    let is_primary = build_runner.is_primary_package(unit);
791    let is_workspace = build_runner.bcx.ws.is_member(&unit.pkg);
792
793    let mut base = build_runner
794        .compilation
795        .rustc_process(unit, is_primary, is_workspace)?;
796    build_base_args(build_runner, &mut base, unit)?;
797    if unit.pkg.manifest().is_embedded() {
798        if !gctx.cli_unstable().script {
799            anyhow::bail!(
800                "parsing `{}` requires `-Zscript`",
801                unit.pkg.manifest_path().display()
802            );
803        }
804        base.arg("-Z").arg("crate-attr=feature(frontmatter)");
805        base.arg("-Z").arg("crate-attr=allow(unused_features)");
806    }
807
808    base.inherit_jobserver(&build_runner.jobserver);
809    build_deps_args(&mut base, build_runner, unit)?;
810    add_cap_lints(build_runner.bcx, unit, &mut base);
811    if let Some(args) = build_runner.bcx.extra_args_for(unit) {
812        base.args(args);
813    }
814    base.args(&unit.rustflags);
815    if gctx.cli_unstable().binary_dep_depinfo {
816        base.arg("-Z").arg("binary-dep-depinfo");
817    }
818    if build_runner.bcx.gctx.cli_unstable().checksum_freshness {
819        base.arg("-Z").arg("checksum-hash-algorithm=blake3");
820    }
821
822    if is_primary {
823        base.env("CARGO_PRIMARY_PACKAGE", "1");
824        let file_list = build_runner.sbom_output_files(unit)?;
825        if !file_list.is_empty() {
826            let file_list = std::env::join_paths(file_list)?;
827            base.env("CARGO_SBOM_PATH", file_list);
828        }
829    }
830
831    if unit.target.is_test() || unit.target.is_bench() {
832        let tmp = build_runner
833            .files()
834            .layout(unit.kind)
835            .build_dir()
836            .prepare_tmp()?;
837        base.env("CARGO_TARGET_TMPDIR", tmp.display().to_string());
838    }
839
840    Ok(base)
841}
842
843/// Prepares flags and environments we can compute for a `rustdoc` invocation
844/// before the job queue starts compiling any unit.
845///
846/// This builds a static view of the invocation. Flags depending on the
847/// completion of other units will be added later in runtime, such as flags
848/// from build scripts.
849fn prepare_rustdoc(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> CargoResult<ProcessBuilder> {
850    let bcx = build_runner.bcx;
851    // script_metadata is not needed here, it is only for tests.
852    let mut rustdoc = build_runner.compilation.rustdoc_process(unit, None)?;
853    if unit.pkg.manifest().is_embedded() {
854        if !bcx.gctx.cli_unstable().script {
855            anyhow::bail!(
856                "parsing `{}` requires `-Zscript`",
857                unit.pkg.manifest_path().display()
858            );
859        }
860        rustdoc.arg("-Z").arg("crate-attr=feature(frontmatter)");
861        rustdoc.arg("-Z").arg("crate-attr=allow(unused_features)");
862    }
863    rustdoc.inherit_jobserver(&build_runner.jobserver);
864    let crate_name = unit.target.crate_name();
865    rustdoc.arg("--crate-name").arg(&crate_name);
866    add_path_args(bcx.ws, unit, &mut rustdoc);
867    add_cap_lints(bcx, unit, &mut rustdoc);
868
869    unit.kind.add_target_arg(&mut rustdoc);
870    let doc_dir = build_runner.files().output_dir(unit);
871    rustdoc.arg("-o").arg(&doc_dir);
872    rustdoc.args(&features_args(unit));
873    rustdoc.args(&check_cfg_args(unit));
874
875    add_error_format_and_color(build_runner, &mut rustdoc);
876    add_allow_features(build_runner, &mut rustdoc);
877
878    if build_runner.bcx.gctx.cli_unstable().rustdoc_depinfo {
879        // toolchain-shared-resources is required for keeping the shared styling resources
880        // invocation-specific is required for keeping the original rustdoc emission
881        let mut arg = if build_runner.bcx.gctx.cli_unstable().rustdoc_mergeable_info {
882            // toolchain resources are written at the end, at the same time as merging
883            OsString::from("--emit=invocation-specific,dep-info=")
884        } else {
885            // if not using mergeable CCI, everything is written every time
886            OsString::from("--emit=toolchain-shared-resources,invocation-specific,dep-info=")
887        };
888        arg.push(rustdoc_dep_info_loc(build_runner, unit));
889        rustdoc.arg(arg);
890
891        if build_runner.bcx.gctx.cli_unstable().checksum_freshness {
892            rustdoc.arg("-Z").arg("checksum-hash-algorithm=blake3");
893        }
894
895        rustdoc.arg("-Zunstable-options");
896    } else if build_runner.bcx.gctx.cli_unstable().rustdoc_mergeable_info {
897        // toolchain resources are written at the end, at the same time as merging
898        rustdoc.arg("--emit=invocation-specific");
899        rustdoc.arg("-Zunstable-options");
900    }
901
902    if build_runner.bcx.gctx.cli_unstable().rustdoc_mergeable_info {
903        // write out mergeable data to be imported
904        rustdoc.arg("--merge=none");
905        let mut arg = OsString::from("--parts-out-dir=");
906        // `-Zrustdoc-mergeable-info` always uses the new layout.
907        arg.push(build_runner.files().out_dir_new_layout(unit));
908        rustdoc.arg(arg);
909    }
910
911    if let Some(trim_paths) = unit.profile.trim_paths.as_ref() {
912        trim_paths_args_rustdoc(&mut rustdoc, build_runner, unit, trim_paths)?;
913    }
914
915    rustdoc.args(unit.pkg.manifest().lint_rustflags());
916
917    let metadata = build_runner.metadata_for_doc_units[unit];
918    rustdoc
919        .arg("-C")
920        .arg(format!("metadata={}", metadata.c_metadata()));
921
922    if unit.mode.is_doc_scrape() {
923        debug_assert!(build_runner.bcx.scrape_units.contains(unit));
924
925        if unit.target.is_test() {
926            rustdoc.arg("--scrape-tests");
927        }
928
929        rustdoc.arg("-Zunstable-options");
930
931        rustdoc
932            .arg("--scrape-examples-output-path")
933            .arg(scrape_output_path(build_runner, unit)?);
934
935        // Only scrape example for items from crates in the workspace, to reduce generated file size
936        for pkg in build_runner.bcx.packages.packages() {
937            let names = pkg
938                .targets()
939                .iter()
940                .map(|target| target.crate_name())
941                .collect::<HashSet<_>>();
942            for name in names {
943                rustdoc.arg("--scrape-examples-target-crate").arg(name);
944            }
945        }
946    }
947
948    if should_include_scrape_units(build_runner.bcx, unit) {
949        rustdoc.arg("-Zunstable-options");
950    }
951
952    build_deps_args(&mut rustdoc, build_runner, unit)?;
953    rustdoc::add_root_urls(build_runner, unit, &mut rustdoc)?;
954
955    rustdoc::add_output_format(build_runner, &mut rustdoc)?;
956
957    if let Some(args) = build_runner.bcx.extra_args_for(unit) {
958        rustdoc.args(args);
959    }
960    rustdoc.args(&unit.rustdocflags);
961
962    if !crate_version_flag_already_present(&rustdoc) {
963        append_crate_version_flag(unit, &mut rustdoc);
964    }
965
966    Ok(rustdoc)
967}
968
969/// Creates a unit of work invoking `rustdoc` for documenting the `unit`.
970fn rustdoc(build_runner: &mut BuildRunner<'_, '_>, unit: &Unit) -> CargoResult<Work> {
971    let mut rustdoc = prepare_rustdoc(build_runner, unit)?;
972
973    let crate_name = unit.target.crate_name();
974    let doc_dir = build_runner.files().output_dir(unit);
975    // Create the documentation directory ahead of time as rustdoc currently has
976    // a bug where concurrent invocations will race to create this directory if
977    // it doesn't already exist.
978    paths::create_dir_all(&doc_dir)?;
979
980    let target_desc = unit.target.description_named();
981    let name = unit.pkg.name();
982    let build_script_outputs = Arc::clone(&build_runner.build_script_outputs);
983    let package_id = unit.pkg.package_id();
984    let target = Target::clone(&unit.target);
985    let manifest = ManifestErrorContext::new(build_runner, unit);
986
987    let rustdoc_dep_info_loc = rustdoc_dep_info_loc(build_runner, unit);
988    let dep_info_loc = fingerprint::dep_info_loc(build_runner, unit);
989    let build_dir = build_runner.bcx.ws.build_dir().into_path_unlocked();
990    let pkg_root = unit.pkg.root().to_path_buf();
991    let cwd = rustdoc
992        .get_cwd()
993        .unwrap_or_else(|| build_runner.bcx.gctx.cwd())
994        .to_path_buf();
995    let fingerprint_dir = build_runner.files().fingerprint_dir(unit);
996    let is_local = unit.is_local();
997    let env_config = Arc::clone(build_runner.bcx.gctx.env_config()?);
998    let rustdoc_depinfo_enabled = build_runner.bcx.gctx.cli_unstable().rustdoc_depinfo;
999
1000    let mut output_options = OutputOptions::new(build_runner, unit);
1001    let script_metadatas = build_runner.find_build_script_metadatas(unit);
1002    let scrape_outputs = if should_include_scrape_units(build_runner.bcx, unit) {
1003        Some(
1004            build_runner
1005                .bcx
1006                .scrape_units
1007                .iter()
1008                .map(|unit| {
1009                    Ok((
1010                        build_runner.files().metadata(unit).unit_id(),
1011                        scrape_output_path(build_runner, unit)?,
1012                    ))
1013                })
1014                .collect::<CargoResult<HashMap<_, _>>>()?,
1015        )
1016    } else {
1017        None
1018    };
1019
1020    let failed_scrape_units = Arc::clone(&build_runner.failed_scrape_units);
1021    let hide_diagnostics_for_scrape_unit = build_runner.bcx.unit_can_fail_for_docscraping(unit)
1022        && !matches!(
1023            build_runner.bcx.gctx.shell().verbosity(),
1024            Verbosity::Verbose
1025        );
1026    let failed_scrape_diagnostic = hide_diagnostics_for_scrape_unit.then(|| {
1027        make_failed_scrape_diagnostic(
1028            build_runner,
1029            unit,
1030            format_args!("failed to scan {target_desc} in package `{name}` for example code usage"),
1031        )
1032    });
1033    if hide_diagnostics_for_scrape_unit {
1034        output_options.show_diagnostics = false;
1035    }
1036
1037    Ok(Work::new(move |state| {
1038        add_custom_flags(
1039            &mut rustdoc,
1040            &build_script_outputs.lock().unwrap(),
1041            script_metadatas,
1042        )?;
1043
1044        // Add the output of scraped examples to the rustdoc command.
1045        // This action must happen after the unit's dependencies have finished,
1046        // because some of those deps may be Docscrape units which have failed.
1047        // So we dynamically determine which `--with-examples` flags to pass here.
1048        if let Some(scrape_outputs) = scrape_outputs {
1049            let failed_scrape_units = failed_scrape_units.lock().unwrap();
1050            for (metadata, output_path) in &scrape_outputs {
1051                if !failed_scrape_units.contains(metadata) {
1052                    rustdoc.arg("--with-examples").arg(output_path);
1053                }
1054            }
1055        }
1056
1057        let crate_dir = doc_dir.join(&crate_name);
1058        if crate_dir.exists() {
1059            // Remove output from a previous build. This ensures that stale
1060            // files for removed items are removed.
1061            debug!("removing pre-existing doc directory {:?}", crate_dir);
1062            paths::remove_dir_all(crate_dir)?;
1063        }
1064        state.running(&rustdoc);
1065        let timestamp = paths::set_invocation_time(&fingerprint_dir)?;
1066
1067        let result = rustdoc
1068            .exec_with_streaming(
1069                &mut |line| on_stdout_line(state, line, package_id, &target),
1070                &mut |line| {
1071                    on_stderr_line(
1072                        state,
1073                        line,
1074                        package_id,
1075                        &manifest,
1076                        &target,
1077                        &mut output_options,
1078                    )
1079                },
1080                false,
1081            )
1082            .map_err(verbose_if_simple_exit_code)
1083            .with_context(|| format!("could not document `{}`", name));
1084
1085        if let Err(e) = result {
1086            if let Some(diagnostic) = failed_scrape_diagnostic {
1087                state.warning(diagnostic);
1088            }
1089
1090            return Err(e);
1091        }
1092
1093        if rustdoc_depinfo_enabled && rustdoc_dep_info_loc.exists() {
1094            fingerprint::translate_dep_info(
1095                &rustdoc_dep_info_loc,
1096                &dep_info_loc,
1097                &cwd,
1098                &pkg_root,
1099                &build_dir,
1100                &rustdoc,
1101                // Should we track source file for doc gen?
1102                is_local,
1103                &env_config,
1104            )
1105            .with_context(|| {
1106                internal(format_args!(
1107                    "could not parse/generate dep info at: {}",
1108                    rustdoc_dep_info_loc.display()
1109                ))
1110            })?;
1111            // This mtime shift allows Cargo to detect if a source file was
1112            // modified in the middle of the build.
1113            paths::set_file_time_no_err(dep_info_loc, timestamp);
1114        }
1115
1116        Ok(())
1117    }))
1118}
1119
1120// The --crate-version flag could have already been passed in RUSTDOCFLAGS
1121// or as an extra compiler argument for rustdoc
1122fn crate_version_flag_already_present(rustdoc: &ProcessBuilder) -> bool {
1123    rustdoc.get_args().any(|flag| {
1124        flag.to_str()
1125            .map_or(false, |flag| flag.starts_with(RUSTDOC_CRATE_VERSION_FLAG))
1126    })
1127}
1128
1129fn append_crate_version_flag(unit: &Unit, rustdoc: &mut ProcessBuilder) {
1130    rustdoc
1131        .arg(RUSTDOC_CRATE_VERSION_FLAG)
1132        .arg(unit.pkg.version().to_string());
1133}
1134
1135/// Adds [`--cap-lints`] to the command to execute.
1136///
1137/// [`--cap-lints`]: https://doc.rust-lang.org/nightly/rustc/lints/levels.html#capping-lints
1138fn add_cap_lints(bcx: &BuildContext<'_, '_>, unit: &Unit, cmd: &mut ProcessBuilder) {
1139    // If this is an upstream dep we don't want warnings from, turn off all
1140    // lints.
1141    if !unit.show_warnings(bcx.gctx) {
1142        cmd.arg("--cap-lints").arg("allow");
1143
1144    // If this is an upstream dep but we *do* want warnings, make sure that they
1145    // don't fail compilation.
1146    } else if !unit.is_local() {
1147        cmd.arg("--cap-lints").arg("warn");
1148    }
1149}
1150
1151/// Forwards [`-Zallow-features`] if it is set for cargo.
1152///
1153/// [`-Zallow-features`]: https://doc.rust-lang.org/nightly/cargo/reference/unstable.html#allow-features
1154fn add_allow_features(build_runner: &BuildRunner<'_, '_>, cmd: &mut ProcessBuilder) {
1155    if let Some(allow) = &build_runner.bcx.gctx.cli_unstable().allow_features {
1156        use std::fmt::Write;
1157        let mut arg = String::from("-Zallow-features=");
1158        for f in allow {
1159            let _ = write!(&mut arg, "{f},");
1160        }
1161        cmd.arg(arg.trim_end_matches(','));
1162    }
1163}
1164
1165/// Adds [`--error-format`] to the command to execute.
1166///
1167/// Cargo always uses JSON output. This has several benefits, such as being
1168/// easier to parse, handles changing formats (for replaying cached messages),
1169/// ensures atomic output (so messages aren't interleaved), allows for
1170/// intercepting messages like rmeta artifacts, etc. rustc includes a
1171/// "rendered" field in the JSON message with the message properly formatted,
1172/// which Cargo will extract and display to the user.
1173///
1174/// [`--error-format`]: https://doc.rust-lang.org/nightly/rustc/command-line-arguments.html#--error-format-control-how-errors-are-produced
1175fn add_error_format_and_color(build_runner: &BuildRunner<'_, '_>, cmd: &mut ProcessBuilder) {
1176    let enable_timings =
1177        build_runner.bcx.gctx.cli_unstable().section_timings && build_runner.bcx.logger.is_some();
1178    if enable_timings {
1179        cmd.arg("-Zunstable-options");
1180    }
1181
1182    cmd.arg("--error-format=json");
1183    let mut json = String::from("--json=diagnostic-rendered-ansi,artifacts,future-incompat");
1184
1185    if let MessageFormat::Short | MessageFormat::Json { short: true, .. } =
1186        build_runner.bcx.build_config.message_format
1187    {
1188        json.push_str(",diagnostic-short");
1189    } else if build_runner.bcx.gctx.shell().err_unicode()
1190        && build_runner.bcx.gctx.cli_unstable().rustc_unicode
1191    {
1192        json.push_str(",diagnostic-unicode");
1193    }
1194
1195    if enable_timings {
1196        json.push_str(",timings");
1197    }
1198
1199    cmd.arg(json);
1200
1201    let gctx = build_runner.bcx.gctx;
1202    if let Some(width) = gctx.shell().err_width().diagnostic_terminal_width() {
1203        cmd.arg(format!("--diagnostic-width={width}"));
1204    }
1205}
1206
1207/// Adds essential rustc flags and environment variables to the command to execute.
1208fn build_base_args(
1209    build_runner: &BuildRunner<'_, '_>,
1210    cmd: &mut ProcessBuilder,
1211    unit: &Unit,
1212) -> CargoResult<()> {
1213    assert!(!unit.mode.is_run_custom_build());
1214
1215    let bcx = build_runner.bcx;
1216    let Profile {
1217        ref opt_level,
1218        codegen_backend,
1219        codegen_units,
1220        debuginfo,
1221        debug_assertions,
1222        split_debuginfo,
1223        overflow_checks,
1224        rpath,
1225        ref panic,
1226        incremental,
1227        strip,
1228        rustflags: profile_rustflags,
1229        trim_paths,
1230        hint_mostly_unused: profile_hint_mostly_unused,
1231        ..
1232    } = unit.profile.clone();
1233    let hints = unit.pkg.hints().cloned().unwrap_or_default();
1234    let test = unit.mode.is_any_test();
1235
1236    let warn = |msg: &str| {
1237        bcx.gctx.shell().warn(format!(
1238            "{}@{}: {msg}",
1239            unit.pkg.package_id().name(),
1240            unit.pkg.package_id().version()
1241        ))
1242    };
1243    let unit_capped_warn = |msg: &str| {
1244        if unit.show_warnings(bcx.gctx) {
1245            warn(msg)
1246        } else {
1247            Ok(())
1248        }
1249    };
1250
1251    cmd.arg("--crate-name").arg(&unit.target.crate_name());
1252
1253    let edition = unit.target.edition();
1254    edition.cmd_edition_arg(cmd);
1255
1256    add_path_args(bcx.ws, unit, cmd);
1257    add_error_format_and_color(build_runner, cmd);
1258    add_allow_features(build_runner, cmd);
1259
1260    let mut contains_dy_lib = false;
1261    if !test {
1262        for crate_type in &unit.target.rustc_crate_types() {
1263            cmd.arg("--crate-type").arg(crate_type.as_str());
1264            contains_dy_lib |= crate_type == &CrateType::Dylib;
1265        }
1266    }
1267
1268    if unit.mode.is_check() {
1269        cmd.arg("--emit=dep-info,metadata");
1270    } else if build_runner.bcx.gctx.cli_unstable().no_embed_metadata {
1271        // Nightly rustc supports the -Zembed-metadata=no flag, which tells it to avoid including
1272        // full metadata in rlib/dylib artifacts, to save space on disk. In this case, metadata
1273        // will only be stored in .rmeta files.
1274        // When we use this flag, we should also pass --emit=metadata to all artifacts that
1275        // contain useful metadata (rlib/dylib/proc macros), so that a .rmeta file is actually
1276        // generated. If we didn't do this, the full metadata would not get written anywhere.
1277        // However, we do not want to pass --emit=metadata to artifacts that never produce useful
1278        // metadata, such as binaries, because that would just unnecessarily create empty .rmeta
1279        // files on disk.
1280        if unit.benefits_from_no_embed_metadata() {
1281            cmd.arg("--emit=dep-info,metadata,link");
1282            cmd.args(&["-Z", "embed-metadata=no"]);
1283        } else {
1284            cmd.arg("--emit=dep-info,link");
1285        }
1286    } else {
1287        // If we don't use -Zembed-metadata=no, we emit .rmeta files only for rlib outputs.
1288        // This metadata may be used in this session for a pipelined compilation, or it may
1289        // be used in a future Cargo session as part of a pipelined compile.
1290        if !unit.requires_upstream_objects() {
1291            cmd.arg("--emit=dep-info,metadata,link");
1292        } else {
1293            cmd.arg("--emit=dep-info,link");
1294        }
1295    }
1296
1297    let prefer_dynamic = (unit.target.for_host() && !unit.target.is_custom_build())
1298        || (contains_dy_lib && !build_runner.is_primary_package(unit));
1299    if prefer_dynamic {
1300        cmd.arg("-C").arg("prefer-dynamic");
1301    }
1302
1303    if opt_level.as_str() != "0" {
1304        cmd.arg("-C").arg(&format!("opt-level={}", opt_level));
1305    }
1306
1307    if *panic != PanicStrategy::Unwind {
1308        cmd.arg("-C").arg(format!("panic={}", panic));
1309    }
1310    if *panic == PanicStrategy::ImmediateAbort {
1311        cmd.arg("-Z").arg("unstable-options");
1312    }
1313
1314    cmd.args(&lto_args(build_runner, unit));
1315
1316    if let Some(backend) = codegen_backend {
1317        cmd.arg("-Z").arg(&format!("codegen-backend={}", backend));
1318    }
1319
1320    if let Some(n) = codegen_units {
1321        cmd.arg("-C").arg(&format!("codegen-units={}", n));
1322    }
1323
1324    let debuginfo = debuginfo.into_inner();
1325    // Shorten the number of arguments if possible.
1326    if debuginfo != TomlDebugInfo::None {
1327        cmd.arg("-C").arg(format!("debuginfo={debuginfo}"));
1328        // This is generally just an optimization on build time so if we don't
1329        // pass it then it's ok. The values for the flag (off, packed, unpacked)
1330        // may be supported or not depending on the platform, so availability is
1331        // checked per-value. For example, at the time of writing this code, on
1332        // Windows the only stable valid value for split-debuginfo is "packed",
1333        // while on Linux "unpacked" is also stable.
1334        if let Some(split) = split_debuginfo {
1335            if build_runner
1336                .bcx
1337                .target_data
1338                .info(unit.kind)
1339                .supports_debuginfo_split(split)
1340            {
1341                cmd.arg("-C").arg(format!("split-debuginfo={split}"));
1342            }
1343        }
1344    }
1345
1346    if let Some(trim_paths) = trim_paths {
1347        trim_paths_args(cmd, build_runner, unit, &trim_paths)?;
1348    }
1349
1350    cmd.args(unit.pkg.manifest().lint_rustflags());
1351    cmd.args(&profile_rustflags);
1352
1353    // `-C overflow-checks` is implied by the setting of `-C debug-assertions`,
1354    // so we only need to provide `-C overflow-checks` if it differs from
1355    // the value of `-C debug-assertions` we would provide.
1356    if opt_level.as_str() != "0" {
1357        if debug_assertions {
1358            cmd.args(&["-C", "debug-assertions=on"]);
1359            if !overflow_checks {
1360                cmd.args(&["-C", "overflow-checks=off"]);
1361            }
1362        } else if overflow_checks {
1363            cmd.args(&["-C", "overflow-checks=on"]);
1364        }
1365    } else if !debug_assertions {
1366        cmd.args(&["-C", "debug-assertions=off"]);
1367        if overflow_checks {
1368            cmd.args(&["-C", "overflow-checks=on"]);
1369        }
1370    } else if !overflow_checks {
1371        cmd.args(&["-C", "overflow-checks=off"]);
1372    }
1373
1374    if test && unit.target.harness() {
1375        cmd.arg("--test");
1376
1377        // Cargo has historically never compiled `--test` binaries with
1378        // `panic=abort` because the `test` crate itself didn't support it.
1379        // Support is now upstream, however, but requires an unstable flag to be
1380        // passed when compiling the test. We require, in Cargo, an unstable
1381        // flag to pass to rustc, so register that here. Eventually this flag
1382        // will simply not be needed when the behavior is stabilized in the Rust
1383        // compiler itself.
1384        if *panic == PanicStrategy::Abort || *panic == PanicStrategy::ImmediateAbort {
1385            cmd.arg("-Z").arg("panic-abort-tests");
1386        }
1387    } else if test {
1388        cmd.arg("--cfg").arg("test");
1389    }
1390
1391    cmd.args(&features_args(unit));
1392    cmd.args(&check_cfg_args(unit));
1393
1394    let meta = build_runner.files().metadata(unit);
1395    cmd.arg("-C")
1396        .arg(&format!("metadata={}", meta.c_metadata()));
1397    if let Some(c_extra_filename) = meta.c_extra_filename() {
1398        cmd.arg("-C")
1399            .arg(&format!("extra-filename=-{c_extra_filename}"));
1400    }
1401
1402    if rpath {
1403        cmd.arg("-C").arg("rpath");
1404    }
1405
1406    cmd.arg("--out-dir")
1407        .arg(&build_runner.files().output_dir(unit));
1408
1409    unit.kind.add_target_arg(cmd);
1410
1411    add_codegen_linker(cmd, build_runner, unit, bcx.gctx.target_applies_to_host()?);
1412
1413    if incremental {
1414        add_codegen_incremental(cmd, build_runner, unit)
1415    }
1416
1417    let pkg_hint_mostly_unused = match hints.mostly_unused {
1418        None => None,
1419        Some(toml::Value::Boolean(b)) => Some(b),
1420        Some(v) => {
1421            unit_capped_warn(&format!(
1422                "ignoring unsupported value type ({}) for 'hints.mostly-unused', which expects a boolean",
1423                v.type_str()
1424            ))?;
1425            None
1426        }
1427    };
1428    if profile_hint_mostly_unused
1429        .or(pkg_hint_mostly_unused)
1430        .unwrap_or(false)
1431    {
1432        if bcx.gctx.cli_unstable().profile_hint_mostly_unused {
1433            cmd.arg("-Zhint-mostly-unused");
1434        } else {
1435            if profile_hint_mostly_unused.is_some() {
1436                // Profiles come from the top-level unit, so we don't use `unit_capped_warn` here.
1437                warn(
1438                    "ignoring 'hint-mostly-unused' profile option, pass `-Zprofile-hint-mostly-unused` to enable it",
1439                )?;
1440            } else if pkg_hint_mostly_unused.is_some() {
1441                unit_capped_warn(
1442                    "ignoring 'hints.mostly-unused', pass `-Zprofile-hint-mostly-unused` to enable it",
1443                )?;
1444            }
1445        }
1446    }
1447
1448    let strip = strip.into_inner();
1449    if strip != StripInner::None {
1450        cmd.arg("-C").arg(format!("strip={}", strip));
1451    }
1452
1453    if unit.is_std {
1454        // -Zforce-unstable-if-unmarked prevents the accidental use of
1455        // unstable crates within the sysroot (such as "extern crate libc" or
1456        // any non-public crate in the sysroot).
1457        //
1458        // RUSTC_BOOTSTRAP allows unstable features on stable.
1459        cmd.arg("-Z")
1460            .arg("force-unstable-if-unmarked")
1461            .env("RUSTC_BOOTSTRAP", "1");
1462    }
1463
1464    Ok(())
1465}
1466
1467/// All active features for the unit passed as `--cfg features=<feature-name>`.
1468fn features_args(unit: &Unit) -> Vec<OsString> {
1469    let mut args = Vec::with_capacity(unit.features.len() * 2);
1470
1471    for feat in &unit.features {
1472        args.push(OsString::from("--cfg"));
1473        args.push(OsString::from(format!("feature=\"{}\"", feat)));
1474    }
1475
1476    args
1477}
1478
1479/// Like [`trim_paths_args`] but for rustdoc invocations.
1480fn trim_paths_args_rustdoc(
1481    cmd: &mut ProcessBuilder,
1482    build_runner: &BuildRunner<'_, '_>,
1483    unit: &Unit,
1484    trim_paths: &TomlTrimPaths,
1485) -> CargoResult<()> {
1486    match trim_paths {
1487        // rustdoc supports diagnostics trimming only.
1488        TomlTrimPaths::Values(values) if !values.contains(&TomlTrimPathsValue::Diagnostics) => {
1489            return Ok(());
1490        }
1491        _ => {}
1492    }
1493
1494    // feature gate was checked during manifest/config parsing.
1495    cmd.arg("-Zunstable-options");
1496
1497    // Order of `--remap-path-prefix` flags is important for `-Zbuild-std`.
1498    // We want to show `/rustc/<hash>/library/std` instead of `std-0.0.0`.
1499    cmd.arg(package_remap(build_runner, unit));
1500    cmd.arg(build_dir_remap(build_runner));
1501    cmd.arg(sysroot_remap(build_runner, unit));
1502
1503    Ok(())
1504}
1505
1506/// Generates the `--remap-path-scope` and `--remap-path-prefix` for [RFC 3127].
1507/// See also unstable feature [`-Ztrim-paths`].
1508///
1509/// [RFC 3127]: https://rust-lang.github.io/rfcs/3127-trim-paths.html
1510/// [`-Ztrim-paths`]: https://doc.rust-lang.org/nightly/cargo/reference/unstable.html#profile-trim-paths-option
1511fn trim_paths_args(
1512    cmd: &mut ProcessBuilder,
1513    build_runner: &BuildRunner<'_, '_>,
1514    unit: &Unit,
1515    trim_paths: &TomlTrimPaths,
1516) -> CargoResult<()> {
1517    if trim_paths.is_none() {
1518        return Ok(());
1519    }
1520
1521    // feature gate was checked during manifest/config parsing.
1522    cmd.arg(format!("--remap-path-scope={trim_paths}"));
1523
1524    // Order of `--remap-path-prefix` flags is important for `-Zbuild-std`.
1525    // We want to show `/rustc/<hash>/library/std` instead of `std-0.0.0`.
1526    cmd.arg(package_remap(build_runner, unit));
1527    cmd.arg(build_dir_remap(build_runner));
1528    cmd.arg(sysroot_remap(build_runner, unit));
1529
1530    Ok(())
1531}
1532
1533/// Path prefix remap rules for sysroot.
1534///
1535/// This remap logic aligns with rustc:
1536/// <https://github.com/rust-lang/rust/blob/c2ef3516/src/bootstrap/src/lib.rs#L1113-L1116>
1537fn sysroot_remap(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> OsString {
1538    let mut remap = OsString::from("--remap-path-prefix=");
1539    remap.push({
1540        // See also `detect_sysroot_src_path()`.
1541        let mut sysroot = build_runner.bcx.target_data.info(unit.kind).sysroot.clone();
1542        sysroot.push("lib");
1543        sysroot.push("rustlib");
1544        sysroot.push("src");
1545        sysroot.push("rust");
1546        sysroot
1547    });
1548    remap.push("=");
1549    remap.push("/rustc/");
1550    if let Some(commit_hash) = build_runner.bcx.rustc().commit_hash.as_ref() {
1551        remap.push(commit_hash);
1552    } else {
1553        remap.push(build_runner.bcx.rustc().version.to_string());
1554    }
1555    remap
1556}
1557
1558/// Path prefix remap rules for dependencies.
1559///
1560/// * Git dependencies: remove `~/.cargo/git/checkouts` prefix.
1561/// * Registry dependencies: remove `~/.cargo/registry/src` prefix.
1562/// * Others (e.g. path dependencies):
1563///     * relative paths to workspace root if inside the workspace directory.
1564///     * otherwise remapped to `<pkg>-<version>`.
1565fn package_remap(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> OsString {
1566    let pkg_root = unit.pkg.root();
1567    let ws_root = build_runner.bcx.ws.root();
1568    let mut remap = OsString::from("--remap-path-prefix=");
1569    let source_id = unit.pkg.package_id().source_id();
1570    if source_id.is_git() {
1571        remap.push(
1572            build_runner
1573                .bcx
1574                .gctx
1575                .git_checkouts_path()
1576                .as_path_unlocked(),
1577        );
1578        remap.push("=");
1579    } else if source_id.is_registry() {
1580        remap.push(
1581            build_runner
1582                .bcx
1583                .gctx
1584                .registry_source_path()
1585                .as_path_unlocked(),
1586        );
1587        remap.push("=");
1588    } else if pkg_root.strip_prefix(ws_root).is_ok() {
1589        remap.push(ws_root);
1590        remap.push("=."); // remap to relative rustc work dir explicitly
1591    } else {
1592        remap.push(pkg_root);
1593        remap.push("=");
1594        remap.push(unit.pkg.name());
1595        remap.push("-");
1596        remap.push(unit.pkg.version().to_string());
1597    }
1598    remap
1599}
1600
1601/// Remap all paths pointing to `build.build-dir`,
1602/// i.e., `[BUILD_DIR]/debug/deps/foo-[HASH].dwo` would be remapped to
1603/// `/cargo/build-dir/debug/deps/foo-[HASH].dwo`
1604/// (note the `/cargo/build-dir` prefix).
1605///
1606/// This covers scenarios like:
1607///
1608/// * Build script generated code. For example, a build script may call `file!`
1609///   macros, and the associated crate uses [`include!`] to include the expanded
1610///   [`file!`] macro in-place via the `OUT_DIR` environment.
1611/// * On Linux, `DW_AT_GNU_dwo_name` that contains paths to split debuginfo
1612///   files (dwp and dwo).
1613fn build_dir_remap(build_runner: &BuildRunner<'_, '_>) -> OsString {
1614    let build_dir = build_runner.bcx.ws.build_dir();
1615    let mut remap = OsString::from("--remap-path-prefix=");
1616    remap.push(build_dir.as_path_unlocked());
1617    remap.push("=/cargo/build-dir");
1618    remap
1619}
1620
1621/// Generates the `--check-cfg` arguments for the `unit`.
1622fn check_cfg_args(unit: &Unit) -> Vec<OsString> {
1623    // The routine below generates the --check-cfg arguments. Our goals here are to
1624    // enable the checking of conditionals and pass the list of declared features.
1625    //
1626    // In the simplified case, it would resemble something like this:
1627    //
1628    //   --check-cfg=cfg() --check-cfg=cfg(feature, values(...))
1629    //
1630    // but having `cfg()` is redundant with the second argument (as well-known names
1631    // and values are implicitly enabled when one or more `--check-cfg` argument is
1632    // passed) so we don't emit it and just pass:
1633    //
1634    //   --check-cfg=cfg(feature, values(...))
1635    //
1636    // This way, even if there are no declared features, the config `feature` will
1637    // still be expected, meaning users would get "unexpected value" instead of name.
1638    // This wasn't always the case, see rust-lang#119930 for some details.
1639
1640    let gross_cap_estimation = unit.pkg.summary().features().len() * 7 + 25;
1641    let mut arg_feature = OsString::with_capacity(gross_cap_estimation);
1642
1643    arg_feature.push("cfg(feature, values(");
1644    for (i, feature) in unit.pkg.summary().features().keys().enumerate() {
1645        if i != 0 {
1646            arg_feature.push(", ");
1647        }
1648        arg_feature.push("\"");
1649        arg_feature.push(feature);
1650        arg_feature.push("\"");
1651    }
1652    arg_feature.push("))");
1653
1654    // In addition to the package features, we also include the `test` cfg (since
1655    // compiler-team#785, as to be able to someday apply it conditionally), as well
1656    // the `docsrs` cfg from the docs.rs service.
1657    //
1658    // We include `docsrs` here (in Cargo) instead of rustc, since there is a much closer
1659    // relationship between Cargo and docs.rs than rustc and docs.rs. In particular, all
1660    // users of docs.rs use Cargo, but not all users of rustc (like Rust-for-Linux) use docs.rs.
1661
1662    vec![
1663        OsString::from("--check-cfg"),
1664        OsString::from("cfg(docsrs,test)"),
1665        OsString::from("--check-cfg"),
1666        arg_feature,
1667    ]
1668}
1669
1670/// Adds LTO related codegen flags.
1671fn lto_args(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> Vec<OsString> {
1672    let mut result = Vec::new();
1673    let mut push = |arg: &str| {
1674        result.push(OsString::from("-C"));
1675        result.push(OsString::from(arg));
1676    };
1677    match build_runner.lto[unit] {
1678        lto::Lto::Run(None) => push("lto"),
1679        lto::Lto::Run(Some(s)) => push(&format!("lto={}", s)),
1680        lto::Lto::Off => {
1681            push("lto=off");
1682            push("embed-bitcode=no");
1683        }
1684        lto::Lto::ObjectAndBitcode => {} // this is rustc's default
1685        lto::Lto::OnlyBitcode => push("linker-plugin-lto"),
1686        lto::Lto::OnlyObject => push("embed-bitcode=no"),
1687    }
1688    result
1689}
1690
1691/// Adds dependency-relevant rustc flags and environment variables
1692/// to the command to execute, such as [`-L`] and [`--extern`].
1693///
1694/// [`-L`]: https://doc.rust-lang.org/nightly/rustc/command-line-arguments.html#-l-add-a-directory-to-the-library-search-path
1695/// [`--extern`]: https://doc.rust-lang.org/nightly/rustc/command-line-arguments.html#--extern-specify-where-an-external-library-is-located
1696fn build_deps_args(
1697    cmd: &mut ProcessBuilder,
1698    build_runner: &BuildRunner<'_, '_>,
1699    unit: &Unit,
1700) -> CargoResult<()> {
1701    let bcx = build_runner.bcx;
1702
1703    for arg in lib_search_paths(build_runner, unit)? {
1704        cmd.arg(arg);
1705    }
1706
1707    let deps = build_runner.unit_deps(unit);
1708
1709    // If there is not one linkable target but should, rustc fails later
1710    // on if there is an `extern crate` for it. This may turn into a hard
1711    // error in the future (see PR #4797).
1712    if !deps
1713        .iter()
1714        .any(|dep| !dep.unit.mode.is_doc() && dep.unit.target.is_linkable())
1715    {
1716        if let Some(dep) = deps.iter().find(|dep| {
1717            !dep.unit.mode.is_doc() && dep.unit.target.is_lib() && !dep.unit.artifact.is_true()
1718        }) {
1719            let dep_name = dep.unit.target.crate_name();
1720            let name = unit.target.crate_name();
1721            bcx.gctx.shell().print_report(&[
1722                Level::WARNING.secondary_title(format!("the package `{dep_name}` provides no linkable target"))
1723                    .elements([
1724                        Level::NOTE.message(format!("this might cause `{name}` to fail compilation")),
1725                        Level::NOTE.message("this warning might turn into a hard error in the future"),
1726                        Level::HELP.message(format!("consider adding 'dylib' or 'rlib' to key 'crate-type' in `{dep_name}`'s Cargo.toml"))
1727                    ])
1728            ], false)?;
1729        }
1730    }
1731
1732    let mut unstable_opts = false;
1733
1734    // Add `OUT_DIR` environment variables for build scripts
1735    let first_custom_build_dep = deps.iter().find(|dep| dep.unit.mode.is_run_custom_build());
1736    if let Some(dep) = first_custom_build_dep {
1737        let out_dir = if bcx.gctx.cli_unstable().build_dir_new_layout {
1738            build_runner.files().out_dir_new_layout(&dep.unit)
1739        } else {
1740            build_runner.files().build_script_out_dir(&dep.unit)
1741        };
1742        cmd.env("OUT_DIR", &out_dir);
1743    }
1744
1745    // Adding output directory for each build script
1746    let is_multiple_build_scripts_enabled = unit
1747        .pkg
1748        .manifest()
1749        .unstable_features()
1750        .require(Feature::multiple_build_scripts())
1751        .is_ok();
1752
1753    if is_multiple_build_scripts_enabled {
1754        for dep in deps {
1755            if dep.unit.mode.is_run_custom_build() {
1756                let out_dir = if bcx.gctx.cli_unstable().build_dir_new_layout {
1757                    build_runner.files().out_dir_new_layout(&dep.unit)
1758                } else {
1759                    build_runner.files().build_script_out_dir(&dep.unit)
1760                };
1761                let target_name = dep.unit.target.name();
1762                let out_dir_prefix = target_name
1763                    .strip_prefix("build-script-")
1764                    .unwrap_or(target_name);
1765                let out_dir_name = format!("{out_dir_prefix}_OUT_DIR");
1766                cmd.env(&out_dir_name, &out_dir);
1767            }
1768        }
1769    }
1770    for arg in extern_args(build_runner, unit, &mut unstable_opts)? {
1771        cmd.arg(arg);
1772    }
1773
1774    for (var, env) in artifact::get_env(build_runner, unit, deps)? {
1775        cmd.env(&var, env);
1776    }
1777
1778    // This will only be set if we're already using a feature
1779    // requiring nightly rust
1780    if unstable_opts {
1781        cmd.arg("-Z").arg("unstable-options");
1782    }
1783
1784    Ok(())
1785}
1786
1787fn add_dep_arg<'a, 'b: 'a>(
1788    map: &mut BTreeMap<&'a Unit, PathBuf>,
1789    build_runner: &'b BuildRunner<'b, '_>,
1790    unit: &'a Unit,
1791) {
1792    if map.contains_key(&unit) {
1793        return;
1794    }
1795    map.insert(&unit, build_runner.files().deps_dir(&unit));
1796
1797    for dep in build_runner.unit_deps(unit) {
1798        add_dep_arg(map, build_runner, &dep.unit);
1799    }
1800}
1801
1802/// Adds extra rustc flags and environment variables collected from the output
1803/// of a build-script to the command to execute, include custom environment
1804/// variables and `cfg`.
1805fn add_custom_flags(
1806    cmd: &mut ProcessBuilder,
1807    build_script_outputs: &BuildScriptOutputs,
1808    metadata_vec: Option<Vec<UnitHash>>,
1809) -> CargoResult<()> {
1810    if let Some(metadata_vec) = metadata_vec {
1811        for metadata in metadata_vec {
1812            if let Some(output) = build_script_outputs.get(metadata) {
1813                for cfg in output.cfgs.iter() {
1814                    cmd.arg("--cfg").arg(cfg);
1815                }
1816                for check_cfg in &output.check_cfgs {
1817                    cmd.arg("--check-cfg").arg(check_cfg);
1818                }
1819                for (name, value) in output.env.iter() {
1820                    cmd.env(name, value);
1821                }
1822            }
1823        }
1824    }
1825
1826    Ok(())
1827}
1828
1829/// Generate a list of `-L` arguments
1830pub fn lib_search_paths(
1831    build_runner: &BuildRunner<'_, '_>,
1832    unit: &Unit,
1833) -> CargoResult<Vec<OsString>> {
1834    let mut lib_search_paths = Vec::new();
1835    if build_runner.bcx.gctx.cli_unstable().build_dir_new_layout {
1836        let mut map = BTreeMap::new();
1837
1838        // Recursively add all dependency args to rustc process
1839        add_dep_arg(&mut map, build_runner, unit);
1840
1841        let paths = map.into_iter().map(|(_, path)| path).sorted_unstable();
1842
1843        for path in paths {
1844            let mut deps = OsString::from("dependency=");
1845            deps.push(path);
1846            lib_search_paths.extend(["-L".into(), deps]);
1847        }
1848    } else {
1849        let mut deps = OsString::from("dependency=");
1850        deps.push(build_runner.files().deps_dir(unit));
1851        lib_search_paths.extend(["-L".into(), deps]);
1852    }
1853
1854    // Be sure that the host path is also listed. This'll ensure that proc macro
1855    // dependencies are correctly found (for reexported macros).
1856    if !unit.kind.is_host() {
1857        let mut deps = OsString::from("dependency=");
1858        deps.push(build_runner.files().host_deps(unit));
1859        lib_search_paths.extend(["-L".into(), deps]);
1860    }
1861
1862    Ok(lib_search_paths)
1863}
1864
1865/// Generates a list of `--extern` arguments.
1866pub fn extern_args(
1867    build_runner: &BuildRunner<'_, '_>,
1868    unit: &Unit,
1869    unstable_opts: &mut bool,
1870) -> CargoResult<Vec<OsString>> {
1871    let mut result = Vec::new();
1872    let deps = build_runner.unit_deps(unit);
1873
1874    let no_embed_metadata = build_runner.bcx.gctx.cli_unstable().no_embed_metadata;
1875
1876    // Closure to add one dependency to `result`.
1877    let mut link_to =
1878        |dep: &UnitDep, extern_crate_name: InternedString, noprelude: bool| -> CargoResult<()> {
1879            let mut value = OsString::new();
1880            let mut opts = Vec::new();
1881            let is_public_dependency_enabled = unit
1882                .pkg
1883                .manifest()
1884                .unstable_features()
1885                .require(Feature::public_dependency())
1886                .is_ok()
1887                || build_runner.bcx.gctx.cli_unstable().public_dependency;
1888            if !dep.public && unit.target.is_lib() && is_public_dependency_enabled {
1889                opts.push("priv");
1890                *unstable_opts = true;
1891            }
1892            if noprelude {
1893                opts.push("noprelude");
1894                *unstable_opts = true;
1895            }
1896            if !opts.is_empty() {
1897                value.push(opts.join(","));
1898                value.push(":");
1899            }
1900            value.push(extern_crate_name.as_str());
1901            value.push("=");
1902
1903            let mut pass = |file| {
1904                let mut value = value.clone();
1905                value.push(file);
1906                result.push(OsString::from("--extern"));
1907                result.push(value);
1908            };
1909
1910            let outputs = build_runner.outputs(&dep.unit)?;
1911
1912            if build_runner.only_requires_rmeta(unit, &dep.unit) || dep.unit.mode.is_check() {
1913                // Example: rlib dependency for an rlib, rmeta is all that is required.
1914                let output = outputs
1915                    .iter()
1916                    .find(|output| output.flavor == FileFlavor::Rmeta)
1917                    .expect("failed to find rmeta dep for pipelined dep");
1918                pass(&output.path);
1919            } else {
1920                // Example: a bin needs `rlib` for dependencies, it cannot use rmeta.
1921                for output in outputs.iter() {
1922                    if output.flavor == FileFlavor::Linkable {
1923                        pass(&output.path);
1924                    }
1925                    // If we use -Zembed-metadata=no, we also need to pass the path to the
1926                    // corresponding .rmeta file to the linkable artifact, because the
1927                    // normal dependency (rlib) doesn't contain the full metadata.
1928                    else if no_embed_metadata && output.flavor == FileFlavor::Rmeta {
1929                        pass(&output.path);
1930                    }
1931                }
1932            }
1933            Ok(())
1934        };
1935
1936    for dep in deps {
1937        if dep.unit.target.is_linkable() && !dep.unit.mode.is_doc() {
1938            link_to(dep, dep.extern_crate_name, dep.noprelude)?;
1939        }
1940    }
1941    if unit.target.proc_macro() {
1942        // Automatically import `proc_macro`.
1943        result.push(OsString::from("--extern"));
1944        result.push(OsString::from("proc_macro"));
1945    }
1946
1947    Ok(result)
1948}
1949
1950/// Adds `-C linker=<path>` if specified.
1951fn add_codegen_linker(
1952    cmd: &mut ProcessBuilder,
1953    build_runner: &BuildRunner<'_, '_>,
1954    unit: &Unit,
1955    target_applies_to_host: bool,
1956) {
1957    let linker = if unit.target.for_host() && !target_applies_to_host {
1958        build_runner
1959            .compilation
1960            .host_linker()
1961            .map(|s| s.as_os_str())
1962    } else {
1963        build_runner
1964            .compilation
1965            .target_linker(unit.kind)
1966            .map(|s| s.as_os_str())
1967    };
1968
1969    if let Some(linker) = linker {
1970        let mut arg = OsString::from("linker=");
1971        arg.push(linker);
1972        cmd.arg("-C").arg(arg);
1973    }
1974}
1975
1976/// Adds `-C incremental=<path>`.
1977fn add_codegen_incremental(
1978    cmd: &mut ProcessBuilder,
1979    build_runner: &BuildRunner<'_, '_>,
1980    unit: &Unit,
1981) {
1982    let dir = build_runner.files().incremental_dir(&unit);
1983    let mut arg = OsString::from("incremental=");
1984    arg.push(dir.as_os_str());
1985    cmd.arg("-C").arg(arg);
1986}
1987
1988fn envify(s: &str) -> String {
1989    s.chars()
1990        .flat_map(|c| c.to_uppercase())
1991        .map(|c| if c == '-' { '_' } else { c })
1992        .collect()
1993}
1994
1995/// Configuration of the display of messages emitted by the compiler,
1996/// e.g. diagnostics, warnings, errors, and message caching.
1997struct OutputOptions {
1998    /// What format we're emitting from Cargo itself.
1999    format: MessageFormat,
2000    /// Where to write the JSON messages to support playback later if the unit
2001    /// is fresh. The file is created lazily so that in the normal case, lots
2002    /// of empty files are not created. If this is None, the output will not
2003    /// be cached (such as when replaying cached messages).
2004    cache_cell: Option<(PathBuf, OnceCell<File>)>,
2005    /// If `true`, display any diagnostics.
2006    /// Other types of JSON messages are processed regardless
2007    /// of the value of this flag.
2008    ///
2009    /// This is used primarily for cache replay. If you build with `-vv`, the
2010    /// cache will be filled with diagnostics from dependencies. When the
2011    /// cache is replayed without `-vv`, we don't want to show them.
2012    show_diagnostics: bool,
2013    /// Tracks the number of warnings we've seen so far.
2014    warnings_seen: usize,
2015    /// Tracks the number of errors we've seen so far.
2016    errors_seen: usize,
2017}
2018
2019impl OutputOptions {
2020    fn new(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> OutputOptions {
2021        let path = build_runner.files().message_cache_path(unit);
2022        // Remove old cache, ignore ENOENT, which is the common case.
2023        drop(fs::remove_file(&path));
2024        let cache_cell = Some((path, OnceCell::new()));
2025        let show_diagnostics =
2026            build_runner.bcx.gctx.warning_handling().unwrap_or_default() != WarningHandling::Allow;
2027        OutputOptions {
2028            format: build_runner.bcx.build_config.message_format,
2029            cache_cell,
2030            show_diagnostics,
2031            warnings_seen: 0,
2032            errors_seen: 0,
2033        }
2034    }
2035}
2036
2037/// Cloned and sendable context about the manifest file.
2038///
2039/// Sometimes we enrich rustc's errors with some locations in the manifest file; this
2040/// contains a `Send`-able copy of the manifest information that we need for the
2041/// enriched errors.
2042struct ManifestErrorContext {
2043    /// The path to the manifest.
2044    path: PathBuf,
2045    /// The locations of various spans within the manifest.
2046    spans: Option<toml::Spanned<toml::de::DeTable<'static>>>,
2047    /// The raw manifest contents.
2048    contents: Option<String>,
2049    /// A lookup for all the unambiguous renamings, mapping from the original package
2050    /// name to the renamed one.
2051    rename_table: HashMap<InternedString, InternedString>,
2052    /// A list of targets we're compiling for, to determine which of the `[target.<something>.dependencies]`
2053    /// tables might be of interest.
2054    requested_kinds: Vec<CompileKind>,
2055    /// A list of all the collections of cfg values, one collection for each target, to determine
2056    /// which of the `[target.'cfg(...)'.dependencies]` tables might be of interest.
2057    cfgs: Vec<Vec<Cfg>>,
2058    host_name: InternedString,
2059    /// Cargo's working directory (for printing out a more friendly manifest path).
2060    cwd: PathBuf,
2061    /// Terminal width for formatting diagnostics.
2062    term_width: usize,
2063}
2064
2065fn on_stdout_line(
2066    state: &JobState<'_, '_>,
2067    line: &str,
2068    _package_id: PackageId,
2069    _target: &Target,
2070) -> CargoResult<()> {
2071    state.stdout(line.to_string())?;
2072    Ok(())
2073}
2074
2075fn on_stderr_line(
2076    state: &JobState<'_, '_>,
2077    line: &str,
2078    package_id: PackageId,
2079    manifest: &ManifestErrorContext,
2080    target: &Target,
2081    options: &mut OutputOptions,
2082) -> CargoResult<()> {
2083    if on_stderr_line_inner(state, line, package_id, manifest, target, options)? {
2084        // Check if caching is enabled.
2085        if let Some((path, cell)) = &mut options.cache_cell {
2086            // Cache the output, which will be replayed later when Fresh.
2087            let f = cell.try_borrow_mut_with(|| paths::create(path))?;
2088            debug_assert!(!line.contains('\n'));
2089            f.write_all(line.as_bytes())?;
2090            f.write_all(&[b'\n'])?;
2091        }
2092    }
2093    Ok(())
2094}
2095
2096/// Returns true if the line should be cached.
2097fn on_stderr_line_inner(
2098    state: &JobState<'_, '_>,
2099    line: &str,
2100    package_id: PackageId,
2101    manifest: &ManifestErrorContext,
2102    target: &Target,
2103    options: &mut OutputOptions,
2104) -> CargoResult<bool> {
2105    // We primarily want to use this function to process JSON messages from
2106    // rustc. The compiler should always print one JSON message per line, and
2107    // otherwise it may have other output intermingled (think RUST_LOG or
2108    // something like that), so skip over everything that doesn't look like a
2109    // JSON message.
2110    if !line.starts_with('{') {
2111        state.stderr(line.to_string())?;
2112        return Ok(true);
2113    }
2114
2115    let mut compiler_message: Box<serde_json::value::RawValue> = match serde_json::from_str(line) {
2116        Ok(msg) => msg,
2117
2118        // If the compiler produced a line that started with `{` but it wasn't
2119        // valid JSON, maybe it wasn't JSON in the first place! Forward it along
2120        // to stderr.
2121        Err(e) => {
2122            debug!("failed to parse json: {:?}", e);
2123            state.stderr(line.to_string())?;
2124            return Ok(true);
2125        }
2126    };
2127
2128    let count_diagnostic = |level, options: &mut OutputOptions| {
2129        if level == "warning" {
2130            options.warnings_seen += 1;
2131        } else if level == "error" {
2132            options.errors_seen += 1;
2133        }
2134    };
2135
2136    if let Ok(report) = serde_json::from_str::<FutureIncompatReport>(compiler_message.get()) {
2137        for item in &report.future_incompat_report {
2138            count_diagnostic(&*item.diagnostic.level, options);
2139        }
2140        state.future_incompat_report(report.future_incompat_report);
2141        return Ok(true);
2142    }
2143
2144    let res = serde_json::from_str::<SectionTiming>(compiler_message.get());
2145    if let Ok(timing_record) = res {
2146        state.on_section_timing_emitted(timing_record);
2147        return Ok(false);
2148    }
2149
2150    // Returns `true` if the diagnostic was modified.
2151    let add_pub_in_priv_diagnostic = |diag: &mut String| -> bool {
2152        // We are parsing the compiler diagnostic here, as this information isn't
2153        // currently exposed elsewhere.
2154        // At the time of writing this comment, rustc emits two different
2155        // "exported_private_dependencies" errors:
2156        //  - type `FromPriv` from private dependency 'priv_dep' in public interface
2157        //  - struct `FromPriv` from private dependency 'priv_dep' is re-exported
2158        // This regex matches them both. To see if it needs to be updated, grep the rust
2159        // source for "EXPORTED_PRIVATE_DEPENDENCIES".
2160        static PRIV_DEP_REGEX: LazyLock<Regex> =
2161            LazyLock::new(|| Regex::new("from private dependency '([A-Za-z0-9-_]+)'").unwrap());
2162        if let Some(crate_name) = PRIV_DEP_REGEX.captures(diag).and_then(|m| m.get(1))
2163            && let Some(ref contents) = manifest.contents
2164            && let Some(span) = manifest.find_crate_span(crate_name.as_str())
2165        {
2166            let rel_path = pathdiff::diff_paths(&manifest.path, &manifest.cwd)
2167                .unwrap_or_else(|| manifest.path.clone())
2168                .display()
2169                .to_string();
2170            let report = [Group::with_title(Level::NOTE.secondary_title(format!(
2171                "dependency `{}` declared here",
2172                crate_name.as_str()
2173            )))
2174            .element(
2175                Snippet::source(contents)
2176                    .path(rel_path)
2177                    .annotation(AnnotationKind::Context.span(span)),
2178            )];
2179
2180            let rendered = Renderer::styled()
2181                .term_width(manifest.term_width)
2182                .render(&report);
2183            diag.push_str(&rendered);
2184            diag.push('\n');
2185            return true;
2186        }
2187        false
2188    };
2189
2190    // Depending on what we're emitting from Cargo itself, we figure out what to
2191    // do with this JSON message.
2192    match options.format {
2193        // In the "human" output formats (human/short) or if diagnostic messages
2194        // from rustc aren't being included in the output of Cargo's JSON
2195        // messages then we extract the diagnostic (if present) here and handle
2196        // it ourselves.
2197        MessageFormat::Human
2198        | MessageFormat::Short
2199        | MessageFormat::Json {
2200            render_diagnostics: true,
2201            ..
2202        } => {
2203            #[derive(serde::Deserialize)]
2204            struct CompilerMessage<'a> {
2205                // `rendered` contains escape sequences, which can't be
2206                // zero-copy deserialized by serde_json.
2207                // See https://github.com/serde-rs/json/issues/742
2208                rendered: String,
2209                #[serde(borrow)]
2210                message: Cow<'a, str>,
2211                #[serde(borrow)]
2212                level: Cow<'a, str>,
2213                children: Vec<PartialDiagnostic>,
2214                code: Option<DiagnosticCode>,
2215            }
2216
2217            // A partial rustfix::diagnostics::Diagnostic. We deserialize only a
2218            // subset of the fields because rustc's output can be extremely
2219            // deeply nested JSON in pathological cases involving macro
2220            // expansion. Rustfix's Diagnostic struct is recursive containing a
2221            // field `children: Vec<Self>`, and it can cause deserialization to
2222            // hit serde_json's default recursion limit, or overflow the stack
2223            // if we turn that off. Cargo only cares about the 1 field listed
2224            // here.
2225            #[derive(serde::Deserialize)]
2226            struct PartialDiagnostic {
2227                spans: Vec<PartialDiagnosticSpan>,
2228            }
2229
2230            // A partial rustfix::diagnostics::DiagnosticSpan.
2231            #[derive(serde::Deserialize)]
2232            struct PartialDiagnosticSpan {
2233                suggestion_applicability: Option<Applicability>,
2234            }
2235
2236            #[derive(serde::Deserialize)]
2237            struct DiagnosticCode {
2238                code: String,
2239            }
2240
2241            if let Ok(mut msg) = serde_json::from_str::<CompilerMessage<'_>>(compiler_message.get())
2242            {
2243                if msg.message.starts_with("aborting due to")
2244                    || msg.message.ends_with("warning emitted")
2245                    || msg.message.ends_with("warnings emitted")
2246                {
2247                    // Skip this line; we'll print our own summary at the end.
2248                    return Ok(true);
2249                }
2250                // state.stderr will add a newline
2251                if msg.rendered.ends_with('\n') {
2252                    msg.rendered.pop();
2253                }
2254                let mut rendered = msg.rendered;
2255                if options.show_diagnostics {
2256                    let machine_applicable: bool = msg
2257                        .children
2258                        .iter()
2259                        .map(|child| {
2260                            child
2261                                .spans
2262                                .iter()
2263                                .filter_map(|span| span.suggestion_applicability)
2264                                .any(|app| app == Applicability::MachineApplicable)
2265                        })
2266                        .any(|b| b);
2267                    count_diagnostic(&msg.level, options);
2268                    if msg
2269                        .code
2270                        .as_ref()
2271                        .is_some_and(|c| c.code == "exported_private_dependencies")
2272                        && options.format != MessageFormat::Short
2273                    {
2274                        add_pub_in_priv_diagnostic(&mut rendered);
2275                    }
2276                    let lint = msg.code.is_some();
2277                    state.emit_diag(&msg.level, rendered, lint, machine_applicable)?;
2278                }
2279                return Ok(true);
2280            }
2281        }
2282
2283        MessageFormat::Json { ansi, .. } => {
2284            #[derive(serde::Deserialize, serde::Serialize)]
2285            struct CompilerMessage<'a> {
2286                rendered: String,
2287                #[serde(flatten, borrow)]
2288                other: std::collections::BTreeMap<Cow<'a, str>, serde_json::Value>,
2289                code: Option<DiagnosticCode<'a>>,
2290            }
2291
2292            #[derive(serde::Deserialize, serde::Serialize)]
2293            struct DiagnosticCode<'a> {
2294                code: String,
2295                #[serde(flatten, borrow)]
2296                other: std::collections::BTreeMap<Cow<'a, str>, serde_json::Value>,
2297            }
2298
2299            if let Ok(mut error) =
2300                serde_json::from_str::<CompilerMessage<'_>>(compiler_message.get())
2301            {
2302                let modified_diag = if error
2303                    .code
2304                    .as_ref()
2305                    .is_some_and(|c| c.code == "exported_private_dependencies")
2306                {
2307                    add_pub_in_priv_diagnostic(&mut error.rendered)
2308                } else {
2309                    false
2310                };
2311
2312                // Remove color information from the rendered string if color is not
2313                // enabled. Cargo always asks for ANSI colors from rustc. This allows
2314                // cached replay to enable/disable colors without re-invoking rustc.
2315                if !ansi {
2316                    error.rendered = anstream::adapter::strip_str(&error.rendered).to_string();
2317                }
2318                if !ansi || modified_diag {
2319                    let new_line = serde_json::to_string(&error)?;
2320                    compiler_message = serde_json::value::RawValue::from_string(new_line)?;
2321                }
2322            }
2323        }
2324    }
2325
2326    // We always tell rustc to emit messages about artifacts being produced.
2327    // These messages feed into pipelined compilation, as well as timing
2328    // information.
2329    //
2330    // Look for a matching directive and inform Cargo internally that a
2331    // metadata file has been produced.
2332    #[derive(serde::Deserialize)]
2333    struct ArtifactNotification<'a> {
2334        #[serde(borrow)]
2335        artifact: Cow<'a, str>,
2336    }
2337
2338    if let Ok(artifact) = serde_json::from_str::<ArtifactNotification<'_>>(compiler_message.get()) {
2339        trace!("found directive from rustc: `{}`", artifact.artifact);
2340        if artifact.artifact.ends_with(".rmeta") {
2341            debug!("looks like metadata finished early!");
2342            state.rmeta_produced();
2343        }
2344        return Ok(false);
2345    }
2346
2347    // And failing all that above we should have a legitimate JSON diagnostic
2348    // from the compiler, so wrap it in an external Cargo JSON message
2349    // indicating which package it came from and then emit it.
2350
2351    if !options.show_diagnostics {
2352        return Ok(true);
2353    }
2354
2355    #[derive(serde::Deserialize)]
2356    struct CompilerMessage<'a> {
2357        #[serde(borrow)]
2358        message: Cow<'a, str>,
2359        #[serde(borrow)]
2360        level: Cow<'a, str>,
2361    }
2362
2363    if let Ok(msg) = serde_json::from_str::<CompilerMessage<'_>>(compiler_message.get()) {
2364        if msg.message.starts_with("aborting due to")
2365            || msg.message.ends_with("warning emitted")
2366            || msg.message.ends_with("warnings emitted")
2367        {
2368            // Skip this line; we'll print our own summary at the end.
2369            return Ok(true);
2370        }
2371        count_diagnostic(&msg.level, options);
2372    }
2373
2374    let msg = machine_message::FromCompiler {
2375        package_id: package_id.to_spec(),
2376        manifest_path: &manifest.path,
2377        target,
2378        message: compiler_message,
2379    }
2380    .to_json_string();
2381
2382    // Switch json lines from rustc/rustdoc that appear on stderr to stdout
2383    // instead. We want the stdout of Cargo to always be machine parseable as
2384    // stderr has our colorized human-readable messages.
2385    state.stdout(msg)?;
2386    Ok(true)
2387}
2388
2389impl ManifestErrorContext {
2390    fn new(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> ManifestErrorContext {
2391        let mut duplicates = HashSet::new();
2392        let mut rename_table = HashMap::new();
2393
2394        for dep in build_runner.unit_deps(unit) {
2395            let unrenamed_id = dep.unit.pkg.package_id().name();
2396            if duplicates.contains(&unrenamed_id) {
2397                continue;
2398            }
2399            match rename_table.entry(unrenamed_id) {
2400                std::collections::hash_map::Entry::Occupied(occ) => {
2401                    occ.remove_entry();
2402                    duplicates.insert(unrenamed_id);
2403                }
2404                std::collections::hash_map::Entry::Vacant(vac) => {
2405                    vac.insert(dep.extern_crate_name);
2406                }
2407            }
2408        }
2409
2410        let bcx = build_runner.bcx;
2411        ManifestErrorContext {
2412            path: unit.pkg.manifest_path().to_owned(),
2413            spans: unit.pkg.manifest().document().cloned(),
2414            contents: unit.pkg.manifest().contents().map(String::from),
2415            requested_kinds: bcx.target_data.requested_kinds().to_owned(),
2416            host_name: bcx.rustc().host,
2417            rename_table,
2418            cwd: path_args(build_runner.bcx.ws, unit).1,
2419            cfgs: bcx
2420                .target_data
2421                .requested_kinds()
2422                .iter()
2423                .map(|k| bcx.target_data.cfg(*k).to_owned())
2424                .collect(),
2425            term_width: bcx
2426                .gctx
2427                .shell()
2428                .err_width()
2429                .diagnostic_terminal_width()
2430                .unwrap_or(annotate_snippets::renderer::DEFAULT_TERM_WIDTH),
2431        }
2432    }
2433
2434    fn requested_target_names(&self) -> impl Iterator<Item = &str> {
2435        self.requested_kinds.iter().map(|kind| match kind {
2436            CompileKind::Host => &self.host_name,
2437            CompileKind::Target(target) => target.short_name(),
2438        })
2439    }
2440
2441    /// Find a span for the dependency that specifies this unrenamed crate, if it's unique.
2442    ///
2443    /// rustc diagnostics (at least for public-in-private) mention the un-renamed
2444    /// crate: if you have `foo = { package = "bar" }`, the rustc diagnostic will
2445    /// say "bar".
2446    ///
2447    /// This function does its best to find a span for "bar", but it could fail if
2448    /// there are multiple candidates:
2449    ///
2450    /// ```toml
2451    /// foo = { package = "bar" }
2452    /// baz = { path = "../bar", package = "bar" }
2453    /// ```
2454    fn find_crate_span(&self, unrenamed: &str) -> Option<Range<usize>> {
2455        let Some(ref spans) = self.spans else {
2456            return None;
2457        };
2458
2459        let orig_name = self.rename_table.get(unrenamed)?.as_str();
2460
2461        if let Some((k, v)) = get_key_value(&spans, &["dependencies", orig_name]) {
2462            // We make some effort to find the unrenamed text: in
2463            //
2464            // ```
2465            // foo = { package = "bar" }
2466            // ```
2467            //
2468            // we try to find the "bar", but fall back to "foo" if we can't (which might
2469            // happen if the renaming took place in the workspace, for example).
2470            if let Some(package) = v.get_ref().as_table().and_then(|t| t.get("package")) {
2471                return Some(package.span());
2472            } else {
2473                return Some(k.span());
2474            }
2475        }
2476
2477        // The dependency could also be in a target-specific table, like
2478        // [target.x86_64-unknown-linux-gnu.dependencies] or
2479        // [target.'cfg(something)'.dependencies]. We filter out target tables
2480        // that don't match a requested target or a requested cfg.
2481        if let Some(target) = spans
2482            .as_ref()
2483            .get("target")
2484            .and_then(|t| t.as_ref().as_table())
2485        {
2486            for (platform, platform_table) in target.iter() {
2487                match platform.as_ref().parse::<Platform>() {
2488                    Ok(Platform::Name(name)) => {
2489                        if !self.requested_target_names().any(|n| n == name) {
2490                            continue;
2491                        }
2492                    }
2493                    Ok(Platform::Cfg(cfg_expr)) => {
2494                        if !self.cfgs.iter().any(|cfgs| cfg_expr.matches(cfgs)) {
2495                            continue;
2496                        }
2497                    }
2498                    Err(_) => continue,
2499                }
2500
2501                let Some(platform_table) = platform_table.as_ref().as_table() else {
2502                    continue;
2503                };
2504
2505                if let Some(deps) = platform_table
2506                    .get("dependencies")
2507                    .and_then(|d| d.as_ref().as_table())
2508                {
2509                    if let Some((k, v)) = deps.get_key_value(orig_name) {
2510                        if let Some(package) = v.get_ref().as_table().and_then(|t| t.get("package"))
2511                        {
2512                            return Some(package.span());
2513                        } else {
2514                            return Some(k.span());
2515                        }
2516                    }
2517                }
2518            }
2519        }
2520        None
2521    }
2522}
2523
2524/// Creates a unit of work that replays the cached compiler message.
2525///
2526/// Usually used when a job is fresh and doesn't need to recompile.
2527fn replay_output_cache(
2528    package_id: PackageId,
2529    manifest: ManifestErrorContext,
2530    target: &Target,
2531    path: PathBuf,
2532    format: MessageFormat,
2533    show_diagnostics: bool,
2534) -> Work {
2535    let target = target.clone();
2536    let mut options = OutputOptions {
2537        format,
2538        cache_cell: None,
2539        show_diagnostics,
2540        warnings_seen: 0,
2541        errors_seen: 0,
2542    };
2543    Work::new(move |state| {
2544        if !path.exists() {
2545            // No cached output, probably didn't emit anything.
2546            return Ok(());
2547        }
2548        // We sometimes have gigabytes of output from the compiler, so avoid
2549        // loading it all into memory at once, as that can cause OOM where
2550        // otherwise there would be none.
2551        let file = paths::open(&path)?;
2552        let mut reader = std::io::BufReader::new(file);
2553        let mut line = String::new();
2554        loop {
2555            let length = reader.read_line(&mut line)?;
2556            if length == 0 {
2557                break;
2558            }
2559            let trimmed = line.trim_end_matches(&['\n', '\r'][..]);
2560            on_stderr_line(state, trimmed, package_id, &manifest, &target, &mut options)?;
2561            line.clear();
2562        }
2563        Ok(())
2564    })
2565}
2566
2567/// Provides a package name with descriptive target information,
2568/// e.g., '`foo` (bin "bar" test)', '`foo` (lib doctest)'.
2569fn descriptive_pkg_name(name: &str, target: &Target, mode: &CompileMode) -> String {
2570    let desc_name = target.description_named();
2571    let mode = if mode.is_rustc_test() && !(target.is_test() || target.is_bench()) {
2572        " test"
2573    } else if mode.is_doc_test() {
2574        " doctest"
2575    } else if mode.is_doc() {
2576        " doc"
2577    } else {
2578        ""
2579    };
2580    format!("`{name}` ({desc_name}{mode})")
2581}
2582
2583/// Applies environment variables from config `[env]` to [`ProcessBuilder`].
2584pub(crate) fn apply_env_config(
2585    gctx: &crate::GlobalContext,
2586    cmd: &mut ProcessBuilder,
2587) -> CargoResult<()> {
2588    for (key, value) in gctx.env_config()?.iter() {
2589        // never override a value that has already been set by cargo
2590        if cmd.get_envs().contains_key(key) {
2591            continue;
2592        }
2593        cmd.env(key, value);
2594    }
2595    Ok(())
2596}
2597
2598/// Checks if there are some scrape units waiting to be processed.
2599fn should_include_scrape_units(bcx: &BuildContext<'_, '_>, unit: &Unit) -> bool {
2600    unit.mode.is_doc() && bcx.scrape_units.len() > 0 && bcx.ws.unit_needs_doc_scrape(unit)
2601}
2602
2603/// Gets the file path of function call information output from `rustdoc`.
2604fn scrape_output_path(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> CargoResult<PathBuf> {
2605    assert!(unit.mode.is_doc() || unit.mode.is_doc_scrape());
2606    build_runner
2607        .outputs(unit)
2608        .map(|outputs| outputs[0].path.clone())
2609}
2610
2611/// Gets the dep-info file emitted by rustdoc.
2612fn rustdoc_dep_info_loc(build_runner: &BuildRunner<'_, '_>, unit: &Unit) -> PathBuf {
2613    let mut loc = build_runner.files().fingerprint_file_path(unit, "");
2614    loc.set_extension("d");
2615    loc
2616}