1use std::assert_matches::assert_matches;
6use std::fmt::Formatter;
7
8use rustc_abi::{BackendRepr, FIRST_VARIANT, FieldIdx, Size, VariantIdx};
9use rustc_const_eval::const_eval::{DummyMachine, throw_machine_stop_str};
10use rustc_const_eval::interpret::{
11 ImmTy, Immediate, InterpCx, OpTy, PlaceTy, Projectable, interp_ok,
12};
13use rustc_data_structures::fx::FxHashMap;
14use rustc_hir::def::DefKind;
15use rustc_middle::bug;
16use rustc_middle::mir::interpret::{InterpResult, Scalar};
17use rustc_middle::mir::visit::{MutVisitor, PlaceContext, Visitor};
18use rustc_middle::mir::*;
19use rustc_middle::ty::layout::LayoutOf;
20use rustc_middle::ty::{self, Ty, TyCtxt};
21use rustc_mir_dataflow::fmt::DebugWithContext;
22use rustc_mir_dataflow::lattice::{FlatSet, HasBottom};
23use rustc_mir_dataflow::value_analysis::{
24 Map, PlaceIndex, State, TrackElem, ValueOrPlace, debug_with_context,
25};
26use rustc_mir_dataflow::{Analysis, Results, ResultsVisitor};
27use rustc_span::DUMMY_SP;
28use tracing::{debug, debug_span, instrument};
29
30const BLOCK_LIMIT: usize = 100;
33const PLACE_LIMIT: usize = 100;
34
35pub(super) struct DataflowConstProp;
36
37impl<'tcx> crate::MirPass<'tcx> for DataflowConstProp {
38 fn is_enabled(&self, sess: &rustc_session::Session) -> bool {
39 sess.mir_opt_level() >= 3
40 }
41
42 #[instrument(skip_all level = "debug")]
43 fn run_pass(&self, tcx: TyCtxt<'tcx>, body: &mut Body<'tcx>) {
44 debug!(def_id = ?body.source.def_id());
45 if tcx.sess.mir_opt_level() < 4 && body.basic_blocks.len() > BLOCK_LIMIT {
46 debug!("aborted dataflow const prop due too many basic blocks");
47 return;
48 }
49
50 let place_limit = if tcx.sess.mir_opt_level() < 4 { Some(PLACE_LIMIT) } else { None };
59
60 let map = Map::new(tcx, body, place_limit);
62
63 let analysis = ConstAnalysis::new(tcx, body, map);
65 let mut results =
66 debug_span!("analyze").in_scope(|| analysis.iterate_to_fixpoint(tcx, body, None));
67
68 let mut visitor = Collector::new(tcx, &body.local_decls);
70 debug_span!("collect").in_scope(|| results.visit_reachable_with(body, &mut visitor));
71 let mut patch = visitor.patch;
72 debug_span!("patch").in_scope(|| patch.visit_body_preserves_cfg(body));
73 }
74
75 fn is_required(&self) -> bool {
76 false
77 }
78}
79
80struct ConstAnalysis<'a, 'tcx> {
85 map: Map<'tcx>,
86 tcx: TyCtxt<'tcx>,
87 local_decls: &'a LocalDecls<'tcx>,
88 ecx: InterpCx<'tcx, DummyMachine>,
89 typing_env: ty::TypingEnv<'tcx>,
90}
91
92impl<'tcx> Analysis<'tcx> for ConstAnalysis<'_, 'tcx> {
93 type Domain = State<FlatSet<Scalar>>;
94
95 const NAME: &'static str = "ConstAnalysis";
96
97 fn bottom_value(&self, _body: &Body<'tcx>) -> Self::Domain {
101 State::Unreachable
102 }
103
104 fn initialize_start_block(&self, body: &Body<'tcx>, state: &mut Self::Domain) {
105 assert_matches!(state, State::Unreachable);
107 *state = State::new_reachable();
108 for arg in body.args_iter() {
109 state.flood(PlaceRef { local: arg, projection: &[] }, &self.map);
110 }
111 }
112
113 fn apply_primary_statement_effect(
114 &mut self,
115 state: &mut Self::Domain,
116 statement: &Statement<'tcx>,
117 _location: Location,
118 ) {
119 if state.is_reachable() {
120 self.handle_statement(statement, state);
121 }
122 }
123
124 fn apply_primary_terminator_effect<'mir>(
125 &mut self,
126 state: &mut Self::Domain,
127 terminator: &'mir Terminator<'tcx>,
128 _location: Location,
129 ) -> TerminatorEdges<'mir, 'tcx> {
130 if state.is_reachable() {
131 self.handle_terminator(terminator, state)
132 } else {
133 TerminatorEdges::None
134 }
135 }
136
137 fn apply_call_return_effect(
138 &mut self,
139 state: &mut Self::Domain,
140 _block: BasicBlock,
141 return_places: CallReturnPlaces<'_, 'tcx>,
142 ) {
143 if state.is_reachable() {
144 self.handle_call_return(return_places, state)
145 }
146 }
147}
148
149impl<'a, 'tcx> ConstAnalysis<'a, 'tcx> {
150 fn new(tcx: TyCtxt<'tcx>, body: &'a Body<'tcx>, map: Map<'tcx>) -> Self {
151 let typing_env = body.typing_env(tcx);
152 Self {
153 map,
154 tcx,
155 local_decls: &body.local_decls,
156 ecx: InterpCx::new(tcx, DUMMY_SP, typing_env, DummyMachine),
157 typing_env,
158 }
159 }
160
161 fn handle_statement(&self, statement: &Statement<'tcx>, state: &mut State<FlatSet<Scalar>>) {
162 match &statement.kind {
163 StatementKind::Assign(box (place, rvalue)) => {
164 self.handle_assign(*place, rvalue, state);
165 }
166 StatementKind::SetDiscriminant { box place, variant_index } => {
167 self.handle_set_discriminant(*place, *variant_index, state);
168 }
169 StatementKind::Intrinsic(box intrinsic) => {
170 self.handle_intrinsic(intrinsic);
171 }
172 StatementKind::StorageLive(local) | StatementKind::StorageDead(local) => {
173 state.flood_with(
176 Place::from(*local).as_ref(),
177 &self.map,
178 FlatSet::<Scalar>::BOTTOM,
179 );
180 }
181 StatementKind::Deinit(box place) => {
182 state.flood_with(place.as_ref(), &self.map, FlatSet::<Scalar>::BOTTOM);
184 }
185 StatementKind::Retag(..) => {
186 }
188 StatementKind::ConstEvalCounter
189 | StatementKind::Nop
190 | StatementKind::FakeRead(..)
191 | StatementKind::PlaceMention(..)
192 | StatementKind::Coverage(..)
193 | StatementKind::BackwardIncompatibleDropHint { .. }
194 | StatementKind::AscribeUserType(..) => {}
195 }
196 }
197
198 fn handle_intrinsic(&self, intrinsic: &NonDivergingIntrinsic<'tcx>) {
199 match intrinsic {
200 NonDivergingIntrinsic::Assume(..) => {
201 }
203 NonDivergingIntrinsic::CopyNonOverlapping(CopyNonOverlapping {
204 dst: _,
205 src: _,
206 count: _,
207 }) => {
208 }
210 }
211 }
212
213 fn handle_operand(
214 &self,
215 operand: &Operand<'tcx>,
216 state: &mut State<FlatSet<Scalar>>,
217 ) -> ValueOrPlace<FlatSet<Scalar>> {
218 match operand {
219 Operand::Constant(box constant) => {
220 ValueOrPlace::Value(self.handle_constant(constant, state))
221 }
222 Operand::Copy(place) | Operand::Move(place) => {
223 self.map.find(place.as_ref()).map(ValueOrPlace::Place).unwrap_or(ValueOrPlace::TOP)
226 }
227 }
228 }
229
230 fn handle_terminator<'mir>(
233 &self,
234 terminator: &'mir Terminator<'tcx>,
235 state: &mut State<FlatSet<Scalar>>,
236 ) -> TerminatorEdges<'mir, 'tcx> {
237 match &terminator.kind {
238 TerminatorKind::Call { .. } | TerminatorKind::InlineAsm { .. } => {
239 }
241 TerminatorKind::Drop { place, .. } => {
242 state.flood_with(place.as_ref(), &self.map, FlatSet::<Scalar>::BOTTOM);
243 }
244 TerminatorKind::Yield { .. } => {
245 bug!("encountered disallowed terminator");
247 }
248 TerminatorKind::SwitchInt { discr, targets } => {
249 return self.handle_switch_int(discr, targets, state);
250 }
251 TerminatorKind::TailCall { .. } => {
252 }
255 TerminatorKind::Goto { .. }
256 | TerminatorKind::UnwindResume
257 | TerminatorKind::UnwindTerminate(_)
258 | TerminatorKind::Return
259 | TerminatorKind::Unreachable
260 | TerminatorKind::Assert { .. }
261 | TerminatorKind::CoroutineDrop
262 | TerminatorKind::FalseEdge { .. }
263 | TerminatorKind::FalseUnwind { .. } => {
264 }
266 }
267 terminator.edges()
268 }
269
270 fn handle_call_return(
271 &self,
272 return_places: CallReturnPlaces<'_, 'tcx>,
273 state: &mut State<FlatSet<Scalar>>,
274 ) {
275 return_places.for_each(|place| {
276 state.flood(place.as_ref(), &self.map);
277 })
278 }
279
280 fn handle_set_discriminant(
281 &self,
282 place: Place<'tcx>,
283 variant_index: VariantIdx,
284 state: &mut State<FlatSet<Scalar>>,
285 ) {
286 state.flood_discr(place.as_ref(), &self.map);
287 if self.map.find_discr(place.as_ref()).is_some() {
288 let enum_ty = place.ty(self.local_decls, self.tcx).ty;
289 if let Some(discr) = self.eval_discriminant(enum_ty, variant_index) {
290 state.assign_discr(
291 place.as_ref(),
292 ValueOrPlace::Value(FlatSet::Elem(discr)),
293 &self.map,
294 );
295 }
296 }
297 }
298
299 fn handle_assign(
300 &self,
301 target: Place<'tcx>,
302 rvalue: &Rvalue<'tcx>,
303 state: &mut State<FlatSet<Scalar>>,
304 ) {
305 match rvalue {
306 Rvalue::Use(operand) => {
307 state.flood(target.as_ref(), &self.map);
308 if let Some(target) = self.map.find(target.as_ref()) {
309 self.assign_operand(state, target, operand);
310 }
311 }
312 Rvalue::CopyForDeref(rhs) => {
313 state.flood(target.as_ref(), &self.map);
314 if let Some(target) = self.map.find(target.as_ref()) {
315 self.assign_operand(state, target, &Operand::Copy(*rhs));
316 }
317 }
318 Rvalue::Aggregate(kind, operands) => {
319 state.flood(target.as_ref(), &self.map);
322
323 let Some(target_idx) = self.map.find(target.as_ref()) else { return };
324
325 let (variant_target, variant_index) = match **kind {
326 AggregateKind::Tuple | AggregateKind::Closure(..) => (Some(target_idx), None),
327 AggregateKind::Adt(def_id, variant_index, ..) => {
328 match self.tcx.def_kind(def_id) {
329 DefKind::Struct => (Some(target_idx), None),
330 DefKind::Enum => (
331 self.map.apply(target_idx, TrackElem::Variant(variant_index)),
332 Some(variant_index),
333 ),
334 _ => return,
335 }
336 }
337 _ => return,
338 };
339 if let Some(variant_target_idx) = variant_target {
340 for (field_index, operand) in operands.iter_enumerated() {
341 if let Some(field) =
342 self.map.apply(variant_target_idx, TrackElem::Field(field_index))
343 {
344 self.assign_operand(state, field, operand);
345 }
346 }
347 }
348 if let Some(variant_index) = variant_index
349 && let Some(discr_idx) = self.map.apply(target_idx, TrackElem::Discriminant)
350 {
351 let enum_ty = target.ty(self.local_decls, self.tcx).ty;
357 if let Some(discr_val) = self.eval_discriminant(enum_ty, variant_index) {
358 state.insert_value_idx(discr_idx, FlatSet::Elem(discr_val), &self.map);
359 }
360 }
361 }
362 Rvalue::BinaryOp(op, box (left, right)) if op.is_overflowing() => {
363 state.flood(target.as_ref(), &self.map);
365
366 let Some(target) = self.map.find(target.as_ref()) else { return };
367
368 let value_target = self.map.apply(target, TrackElem::Field(0_u32.into()));
369 let overflow_target = self.map.apply(target, TrackElem::Field(1_u32.into()));
370
371 if value_target.is_some() || overflow_target.is_some() {
372 let (val, overflow) = self.binary_op(state, *op, left, right);
373
374 if let Some(value_target) = value_target {
375 state.insert_value_idx(value_target, val, &self.map);
377 }
378 if let Some(overflow_target) = overflow_target {
379 state.insert_value_idx(overflow_target, overflow, &self.map);
381 }
382 }
383 }
384 Rvalue::Cast(
385 CastKind::PointerCoercion(ty::adjustment::PointerCoercion::Unsize, _),
386 operand,
387 _,
388 ) => {
389 let pointer = self.handle_operand(operand, state);
390 state.assign(target.as_ref(), pointer, &self.map);
391
392 if let Some(target_len) = self.map.find_len(target.as_ref())
393 && let operand_ty = operand.ty(self.local_decls, self.tcx)
394 && let Some(operand_ty) = operand_ty.builtin_deref(true)
395 && let ty::Array(_, len) = operand_ty.kind()
396 && let Some(len) = Const::Ty(self.tcx.types.usize, *len)
397 .try_eval_scalar_int(self.tcx, self.typing_env)
398 {
399 state.insert_value_idx(target_len, FlatSet::Elem(len.into()), &self.map);
400 }
401 }
402 _ => {
403 let result = self.handle_rvalue(rvalue, state);
404 state.assign(target.as_ref(), result, &self.map);
405 }
406 }
407 }
408
409 fn handle_rvalue(
410 &self,
411 rvalue: &Rvalue<'tcx>,
412 state: &mut State<FlatSet<Scalar>>,
413 ) -> ValueOrPlace<FlatSet<Scalar>> {
414 let val = match rvalue {
415 Rvalue::Len(place) => {
416 let place_ty = place.ty(self.local_decls, self.tcx);
417 if let ty::Array(_, len) = place_ty.ty.kind() {
418 Const::Ty(self.tcx.types.usize, *len)
419 .try_eval_scalar(self.tcx, self.typing_env)
420 .map_or(FlatSet::Top, FlatSet::Elem)
421 } else if let [ProjectionElem::Deref] = place.projection[..] {
422 state.get_len(place.local.into(), &self.map)
423 } else {
424 FlatSet::Top
425 }
426 }
427 Rvalue::Cast(CastKind::IntToInt | CastKind::IntToFloat, operand, ty) => {
428 let Ok(layout) = self.tcx.layout_of(self.typing_env.as_query_input(*ty)) else {
429 return ValueOrPlace::Value(FlatSet::Top);
430 };
431 match self.eval_operand(operand, state) {
432 FlatSet::Elem(op) => self
433 .ecx
434 .int_to_int_or_float(&op, layout)
435 .discard_err()
436 .map_or(FlatSet::Top, |result| self.wrap_immediate(*result)),
437 FlatSet::Bottom => FlatSet::Bottom,
438 FlatSet::Top => FlatSet::Top,
439 }
440 }
441 Rvalue::Cast(CastKind::FloatToInt | CastKind::FloatToFloat, operand, ty) => {
442 let Ok(layout) = self.tcx.layout_of(self.typing_env.as_query_input(*ty)) else {
443 return ValueOrPlace::Value(FlatSet::Top);
444 };
445 match self.eval_operand(operand, state) {
446 FlatSet::Elem(op) => self
447 .ecx
448 .float_to_float_or_int(&op, layout)
449 .discard_err()
450 .map_or(FlatSet::Top, |result| self.wrap_immediate(*result)),
451 FlatSet::Bottom => FlatSet::Bottom,
452 FlatSet::Top => FlatSet::Top,
453 }
454 }
455 Rvalue::Cast(CastKind::Transmute, operand, _) => {
456 match self.eval_operand(operand, state) {
457 FlatSet::Elem(op) => self.wrap_immediate(*op),
458 FlatSet::Bottom => FlatSet::Bottom,
459 FlatSet::Top => FlatSet::Top,
460 }
461 }
462 Rvalue::BinaryOp(op, box (left, right)) if !op.is_overflowing() => {
463 let (val, _overflow) = self.binary_op(state, *op, left, right);
466 val
467 }
468 Rvalue::UnaryOp(op, operand) => match self.eval_operand(operand, state) {
469 FlatSet::Elem(value) => self
470 .ecx
471 .unary_op(*op, &value)
472 .discard_err()
473 .map_or(FlatSet::Top, |val| self.wrap_immediate(*val)),
474 FlatSet::Bottom => FlatSet::Bottom,
475 FlatSet::Top => FlatSet::Top,
476 },
477 Rvalue::NullaryOp(null_op, ty) => {
478 let Ok(layout) = self.tcx.layout_of(self.typing_env.as_query_input(*ty)) else {
479 return ValueOrPlace::Value(FlatSet::Top);
480 };
481 let val = match null_op {
482 NullOp::SizeOf if layout.is_sized() => layout.size.bytes(),
483 NullOp::AlignOf if layout.is_sized() => layout.align.abi.bytes(),
484 NullOp::OffsetOf(fields) => self
485 .ecx
486 .tcx
487 .offset_of_subfield(self.typing_env, layout, fields.iter())
488 .bytes(),
489 _ => return ValueOrPlace::Value(FlatSet::Top),
490 };
491 FlatSet::Elem(Scalar::from_target_usize(val, &self.tcx))
492 }
493 Rvalue::Discriminant(place) => state.get_discr(place.as_ref(), &self.map),
494 Rvalue::Use(operand) => return self.handle_operand(operand, state),
495 Rvalue::CopyForDeref(place) => {
496 return self.handle_operand(&Operand::Copy(*place), state);
497 }
498 Rvalue::Ref(..) | Rvalue::RawPtr(..) => {
499 return ValueOrPlace::TOP;
501 }
502 Rvalue::Repeat(..)
503 | Rvalue::ThreadLocalRef(..)
504 | Rvalue::Cast(..)
505 | Rvalue::BinaryOp(..)
506 | Rvalue::Aggregate(..)
507 | Rvalue::ShallowInitBox(..)
508 | Rvalue::WrapUnsafeBinder(..) => {
509 return ValueOrPlace::TOP;
511 }
512 };
513 ValueOrPlace::Value(val)
514 }
515
516 fn handle_constant(
517 &self,
518 constant: &ConstOperand<'tcx>,
519 _state: &mut State<FlatSet<Scalar>>,
520 ) -> FlatSet<Scalar> {
521 constant
522 .const_
523 .try_eval_scalar(self.tcx, self.typing_env)
524 .map_or(FlatSet::Top, FlatSet::Elem)
525 }
526
527 fn handle_switch_int<'mir>(
528 &self,
529 discr: &'mir Operand<'tcx>,
530 targets: &'mir SwitchTargets,
531 state: &mut State<FlatSet<Scalar>>,
532 ) -> TerminatorEdges<'mir, 'tcx> {
533 let value = match self.handle_operand(discr, state) {
534 ValueOrPlace::Value(value) => value,
535 ValueOrPlace::Place(place) => state.get_idx(place, &self.map),
536 };
537 match value {
538 FlatSet::Bottom => TerminatorEdges::None,
541 FlatSet::Elem(scalar) => {
542 if let Ok(scalar_int) = scalar.try_to_scalar_int() {
543 TerminatorEdges::Single(
544 targets.target_for_value(scalar_int.to_bits_unchecked()),
545 )
546 } else {
547 TerminatorEdges::SwitchInt { discr, targets }
548 }
549 }
550 FlatSet::Top => TerminatorEdges::SwitchInt { discr, targets },
551 }
552 }
553
554 fn assign_operand(
556 &self,
557 state: &mut State<FlatSet<Scalar>>,
558 place: PlaceIndex,
559 operand: &Operand<'tcx>,
560 ) {
561 match operand {
562 Operand::Copy(rhs) | Operand::Move(rhs) => {
563 if let Some(rhs) = self.map.find(rhs.as_ref()) {
564 state.insert_place_idx(place, rhs, &self.map);
565 } else if rhs.projection.first() == Some(&PlaceElem::Deref)
566 && let FlatSet::Elem(pointer) = state.get(rhs.local.into(), &self.map)
567 && let rhs_ty = self.local_decls[rhs.local].ty
568 && let Ok(rhs_layout) =
569 self.tcx.layout_of(self.typing_env.as_query_input(rhs_ty))
570 {
571 let op = ImmTy::from_scalar(pointer, rhs_layout).into();
572 self.assign_constant(state, place, op, rhs.projection);
573 }
574 }
575 Operand::Constant(box constant) => {
576 if let Some(constant) =
577 self.ecx.eval_mir_constant(&constant.const_, constant.span, None).discard_err()
578 {
579 self.assign_constant(state, place, constant, &[]);
580 }
581 }
582 }
583 }
584
585 #[instrument(level = "trace", skip(self, state))]
589 fn assign_constant(
590 &self,
591 state: &mut State<FlatSet<Scalar>>,
592 place: PlaceIndex,
593 mut operand: OpTy<'tcx>,
594 projection: &[PlaceElem<'tcx>],
595 ) {
596 for &(mut proj_elem) in projection {
597 if let PlaceElem::Index(index) = proj_elem {
598 if let FlatSet::Elem(index) = state.get(index.into(), &self.map)
599 && let Some(offset) = index.to_target_usize(&self.tcx).discard_err()
600 && let Some(min_length) = offset.checked_add(1)
601 {
602 proj_elem = PlaceElem::ConstantIndex { offset, min_length, from_end: false };
603 } else {
604 return;
605 }
606 }
607 operand = if let Some(operand) = self.ecx.project(&operand, proj_elem).discard_err() {
608 operand
609 } else {
610 return;
611 }
612 }
613
614 self.map.for_each_projection_value(
615 place,
616 operand,
617 &mut |elem, op| match elem {
618 TrackElem::Field(idx) => self.ecx.project_field(op, idx.as_usize()).discard_err(),
619 TrackElem::Variant(idx) => self.ecx.project_downcast(op, idx).discard_err(),
620 TrackElem::Discriminant => {
621 let variant = self.ecx.read_discriminant(op).discard_err()?;
622 let discr_value =
623 self.ecx.discriminant_for_variant(op.layout.ty, variant).discard_err()?;
624 Some(discr_value.into())
625 }
626 TrackElem::DerefLen => {
627 let op: OpTy<'_> = self.ecx.deref_pointer(op).discard_err()?.into();
628 let len_usize = op.len(&self.ecx).discard_err()?;
629 let layout = self
630 .tcx
631 .layout_of(self.typing_env.as_query_input(self.tcx.types.usize))
632 .unwrap();
633 Some(ImmTy::from_uint(len_usize, layout).into())
634 }
635 },
636 &mut |place, op| {
637 if let Some(imm) = self.ecx.read_immediate_raw(op).discard_err()
638 && let Some(imm) = imm.right()
639 {
640 let elem = self.wrap_immediate(*imm);
641 state.insert_value_idx(place, elem, &self.map);
642 }
643 },
644 );
645 }
646
647 fn binary_op(
648 &self,
649 state: &mut State<FlatSet<Scalar>>,
650 op: BinOp,
651 left: &Operand<'tcx>,
652 right: &Operand<'tcx>,
653 ) -> (FlatSet<Scalar>, FlatSet<Scalar>) {
654 let left = self.eval_operand(left, state);
655 let right = self.eval_operand(right, state);
656
657 match (left, right) {
658 (FlatSet::Bottom, _) | (_, FlatSet::Bottom) => (FlatSet::Bottom, FlatSet::Bottom),
659 (FlatSet::Elem(left), FlatSet::Elem(right)) => {
661 match self.ecx.binary_op(op, &left, &right).discard_err() {
662 Some(val) => {
666 if matches!(val.layout.backend_repr, BackendRepr::ScalarPair(..)) {
667 let (val, overflow) = val.to_scalar_pair();
668 (FlatSet::Elem(val), FlatSet::Elem(overflow))
669 } else {
670 (FlatSet::Elem(val.to_scalar()), FlatSet::Bottom)
671 }
672 }
673 _ => (FlatSet::Top, FlatSet::Top),
674 }
675 }
676 (FlatSet::Elem(const_arg), _) | (_, FlatSet::Elem(const_arg)) => {
678 let layout = const_arg.layout;
679 if !matches!(layout.backend_repr, rustc_abi::BackendRepr::Scalar(..)) {
680 return (FlatSet::Top, FlatSet::Top);
681 }
682
683 let arg_scalar = const_arg.to_scalar();
684 let Some(arg_value) = arg_scalar.to_bits(layout.size).discard_err() else {
685 return (FlatSet::Top, FlatSet::Top);
686 };
687
688 match op {
689 BinOp::BitAnd if arg_value == 0 => (FlatSet::Elem(arg_scalar), FlatSet::Bottom),
690 BinOp::BitOr
691 if arg_value == layout.size.truncate(u128::MAX)
692 || (layout.ty.is_bool() && arg_value == 1) =>
693 {
694 (FlatSet::Elem(arg_scalar), FlatSet::Bottom)
695 }
696 BinOp::Mul if layout.ty.is_integral() && arg_value == 0 => {
697 (FlatSet::Elem(arg_scalar), FlatSet::Elem(Scalar::from_bool(false)))
698 }
699 _ => (FlatSet::Top, FlatSet::Top),
700 }
701 }
702 (FlatSet::Top, FlatSet::Top) => (FlatSet::Top, FlatSet::Top),
703 }
704 }
705
706 fn eval_operand(
707 &self,
708 op: &Operand<'tcx>,
709 state: &mut State<FlatSet<Scalar>>,
710 ) -> FlatSet<ImmTy<'tcx>> {
711 let value = match self.handle_operand(op, state) {
712 ValueOrPlace::Value(value) => value,
713 ValueOrPlace::Place(place) => state.get_idx(place, &self.map),
714 };
715 match value {
716 FlatSet::Top => FlatSet::Top,
717 FlatSet::Elem(scalar) => {
718 let ty = op.ty(self.local_decls, self.tcx);
719 self.tcx
720 .layout_of(self.typing_env.as_query_input(ty))
721 .map_or(FlatSet::Top, |layout| {
722 FlatSet::Elem(ImmTy::from_scalar(scalar, layout))
723 })
724 }
725 FlatSet::Bottom => FlatSet::Bottom,
726 }
727 }
728
729 fn eval_discriminant(&self, enum_ty: Ty<'tcx>, variant_index: VariantIdx) -> Option<Scalar> {
730 if !enum_ty.is_enum() {
731 return None;
732 }
733 let enum_ty_layout = self.tcx.layout_of(self.typing_env.as_query_input(enum_ty)).ok()?;
734 let discr_value =
735 self.ecx.discriminant_for_variant(enum_ty_layout.ty, variant_index).discard_err()?;
736 Some(discr_value.to_scalar())
737 }
738
739 fn wrap_immediate(&self, imm: Immediate) -> FlatSet<Scalar> {
740 match imm {
741 Immediate::Scalar(scalar) => FlatSet::Elem(scalar),
742 Immediate::Uninit => FlatSet::Bottom,
743 _ => FlatSet::Top,
744 }
745 }
746}
747
748impl<'tcx> DebugWithContext<ConstAnalysis<'_, 'tcx>> for State<FlatSet<Scalar>> {
750 fn fmt_with(&self, ctxt: &ConstAnalysis<'_, 'tcx>, f: &mut Formatter<'_>) -> std::fmt::Result {
751 match self {
752 State::Reachable(values) => debug_with_context(values, None, &ctxt.map, f),
753 State::Unreachable => write!(f, "unreachable"),
754 }
755 }
756
757 fn fmt_diff_with(
758 &self,
759 old: &Self,
760 ctxt: &ConstAnalysis<'_, 'tcx>,
761 f: &mut Formatter<'_>,
762 ) -> std::fmt::Result {
763 match (self, old) {
764 (State::Reachable(this), State::Reachable(old)) => {
765 debug_with_context(this, Some(old), &ctxt.map, f)
766 }
767 _ => Ok(()), }
769 }
770}
771
772struct Patch<'tcx> {
773 tcx: TyCtxt<'tcx>,
774
775 before_effect: FxHashMap<(Location, Place<'tcx>), Const<'tcx>>,
779
780 assignments: FxHashMap<Location, Const<'tcx>>,
782}
783
784impl<'tcx> Patch<'tcx> {
785 pub(crate) fn new(tcx: TyCtxt<'tcx>) -> Self {
786 Self { tcx, before_effect: FxHashMap::default(), assignments: FxHashMap::default() }
787 }
788
789 fn make_operand(&self, const_: Const<'tcx>) -> Operand<'tcx> {
790 Operand::Constant(Box::new(ConstOperand { span: DUMMY_SP, user_ty: None, const_ }))
791 }
792}
793
794struct Collector<'a, 'tcx> {
795 patch: Patch<'tcx>,
796 local_decls: &'a LocalDecls<'tcx>,
797}
798
799impl<'a, 'tcx> Collector<'a, 'tcx> {
800 pub(crate) fn new(tcx: TyCtxt<'tcx>, local_decls: &'a LocalDecls<'tcx>) -> Self {
801 Self { patch: Patch::new(tcx), local_decls }
802 }
803
804 #[instrument(level = "trace", skip(self, ecx, map), ret)]
805 fn try_make_constant(
806 &self,
807 ecx: &mut InterpCx<'tcx, DummyMachine>,
808 place: Place<'tcx>,
809 state: &State<FlatSet<Scalar>>,
810 map: &Map<'tcx>,
811 ) -> Option<Const<'tcx>> {
812 let ty = place.ty(self.local_decls, self.patch.tcx).ty;
813 let layout = ecx.layout_of(ty).ok()?;
814
815 if layout.is_zst() {
816 return Some(Const::zero_sized(ty));
817 }
818
819 if layout.is_unsized() {
820 return None;
821 }
822
823 let place = map.find(place.as_ref())?;
824 if layout.backend_repr.is_scalar()
825 && let Some(value) = propagatable_scalar(place, state, map)
826 {
827 return Some(Const::Val(ConstValue::Scalar(value), ty));
828 }
829
830 if matches!(layout.backend_repr, BackendRepr::Scalar(..) | BackendRepr::ScalarPair(..)) {
831 let alloc_id = ecx
832 .intern_with_temp_alloc(layout, |ecx, dest| {
833 try_write_constant(ecx, dest, place, ty, state, map)
834 })
835 .discard_err()?;
836 return Some(Const::Val(ConstValue::Indirect { alloc_id, offset: Size::ZERO }, ty));
837 }
838
839 None
840 }
841}
842
843#[instrument(level = "trace", skip(map), ret)]
844fn propagatable_scalar(
845 place: PlaceIndex,
846 state: &State<FlatSet<Scalar>>,
847 map: &Map<'_>,
848) -> Option<Scalar> {
849 if let FlatSet::Elem(value) = state.get_idx(place, map)
850 && value.try_to_scalar_int().is_ok()
851 {
852 Some(value)
854 } else {
855 None
856 }
857}
858
859#[instrument(level = "trace", skip(ecx, state, map), ret)]
860fn try_write_constant<'tcx>(
861 ecx: &mut InterpCx<'tcx, DummyMachine>,
862 dest: &PlaceTy<'tcx>,
863 place: PlaceIndex,
864 ty: Ty<'tcx>,
865 state: &State<FlatSet<Scalar>>,
866 map: &Map<'tcx>,
867) -> InterpResult<'tcx> {
868 let layout = ecx.layout_of(ty)?;
869
870 if layout.is_zst() {
872 return interp_ok(());
873 }
874
875 if layout.backend_repr.is_scalar()
877 && let Some(value) = propagatable_scalar(place, state, map)
878 {
879 return ecx.write_immediate(Immediate::Scalar(value), dest);
880 }
881
882 match ty.kind() {
883 ty::FnDef(..) => {}
885
886 ty::Bool | ty::Int(_) | ty::Uint(_) | ty::Float(_) | ty::Char =>
888 throw_machine_stop_str!("primitive type with provenance"),
889
890 ty::Tuple(elem_tys) => {
891 for (i, elem) in elem_tys.iter().enumerate() {
892 let Some(field) = map.apply(place, TrackElem::Field(FieldIdx::from_usize(i))) else {
893 throw_machine_stop_str!("missing field in tuple")
894 };
895 let field_dest = ecx.project_field(dest, i)?;
896 try_write_constant(ecx, &field_dest, field, elem, state, map)?;
897 }
898 }
899
900 ty::Adt(def, args) => {
901 if def.is_union() {
902 throw_machine_stop_str!("cannot propagate unions")
903 }
904
905 let (variant_idx, variant_def, variant_place, variant_dest) = if def.is_enum() {
906 let Some(discr) = map.apply(place, TrackElem::Discriminant) else {
907 throw_machine_stop_str!("missing discriminant for enum")
908 };
909 let FlatSet::Elem(Scalar::Int(discr)) = state.get_idx(discr, map) else {
910 throw_machine_stop_str!("discriminant with provenance")
911 };
912 let discr_bits = discr.to_bits(discr.size());
913 let Some((variant, _)) = def.discriminants(*ecx.tcx).find(|(_, var)| discr_bits == var.val) else {
914 throw_machine_stop_str!("illegal discriminant for enum")
915 };
916 let Some(variant_place) = map.apply(place, TrackElem::Variant(variant)) else {
917 throw_machine_stop_str!("missing variant for enum")
918 };
919 let variant_dest = ecx.project_downcast(dest, variant)?;
920 (variant, def.variant(variant), variant_place, variant_dest)
921 } else {
922 (FIRST_VARIANT, def.non_enum_variant(), place, dest.clone())
923 };
924
925 for (i, field) in variant_def.fields.iter_enumerated() {
926 let ty = field.ty(*ecx.tcx, args);
927 let Some(field) = map.apply(variant_place, TrackElem::Field(i)) else {
928 throw_machine_stop_str!("missing field in ADT")
929 };
930 let field_dest = ecx.project_field(&variant_dest, i.as_usize())?;
931 try_write_constant(ecx, &field_dest, field, ty, state, map)?;
932 }
933 ecx.write_discriminant(variant_idx, dest)?;
934 }
935
936 ty::Array(_, _)
938 | ty::Pat(_, _)
939
940 | ty::Ref(..) | ty::RawPtr(..) | ty::FnPtr(..) | ty::Str | ty::Slice(_)
942
943 | ty::Never
944 | ty::Foreign(..)
945 | ty::Alias(..)
946 | ty::Param(_)
947 | ty::Bound(..)
948 | ty::Placeholder(..)
949 | ty::Closure(..)
950 | ty::CoroutineClosure(..)
951 | ty::Coroutine(..)
952 | ty::Dynamic(..)
953 | ty::UnsafeBinder(_) => throw_machine_stop_str!("unsupported type"),
954
955 ty::Error(_) | ty::Infer(..) | ty::CoroutineWitness(..) => bug!(),
956 }
957
958 interp_ok(())
959}
960
961impl<'mir, 'tcx> ResultsVisitor<'mir, 'tcx, ConstAnalysis<'_, 'tcx>> for Collector<'_, 'tcx> {
962 #[instrument(level = "trace", skip(self, results, statement))]
963 fn visit_after_early_statement_effect(
964 &mut self,
965 results: &mut Results<'tcx, ConstAnalysis<'_, 'tcx>>,
966 state: &State<FlatSet<Scalar>>,
967 statement: &'mir Statement<'tcx>,
968 location: Location,
969 ) {
970 match &statement.kind {
971 StatementKind::Assign(box (_, rvalue)) => {
972 OperandCollector {
973 state,
974 visitor: self,
975 ecx: &mut results.analysis.ecx,
976 map: &results.analysis.map,
977 }
978 .visit_rvalue(rvalue, location);
979 }
980 _ => (),
981 }
982 }
983
984 #[instrument(level = "trace", skip(self, results, statement))]
985 fn visit_after_primary_statement_effect(
986 &mut self,
987 results: &mut Results<'tcx, ConstAnalysis<'_, 'tcx>>,
988 state: &State<FlatSet<Scalar>>,
989 statement: &'mir Statement<'tcx>,
990 location: Location,
991 ) {
992 match statement.kind {
993 StatementKind::Assign(box (_, Rvalue::Use(Operand::Constant(_)))) => {
994 }
996 StatementKind::Assign(box (place, _)) => {
997 if let Some(value) = self.try_make_constant(
998 &mut results.analysis.ecx,
999 place,
1000 state,
1001 &results.analysis.map,
1002 ) {
1003 self.patch.assignments.insert(location, value);
1004 }
1005 }
1006 _ => (),
1007 }
1008 }
1009
1010 fn visit_after_early_terminator_effect(
1011 &mut self,
1012 results: &mut Results<'tcx, ConstAnalysis<'_, 'tcx>>,
1013 state: &State<FlatSet<Scalar>>,
1014 terminator: &'mir Terminator<'tcx>,
1015 location: Location,
1016 ) {
1017 OperandCollector {
1018 state,
1019 visitor: self,
1020 ecx: &mut results.analysis.ecx,
1021 map: &results.analysis.map,
1022 }
1023 .visit_terminator(terminator, location);
1024 }
1025}
1026
1027impl<'tcx> MutVisitor<'tcx> for Patch<'tcx> {
1028 fn tcx(&self) -> TyCtxt<'tcx> {
1029 self.tcx
1030 }
1031
1032 fn visit_statement(&mut self, statement: &mut Statement<'tcx>, location: Location) {
1033 if let Some(value) = self.assignments.get(&location) {
1034 match &mut statement.kind {
1035 StatementKind::Assign(box (_, rvalue)) => {
1036 *rvalue = Rvalue::Use(self.make_operand(*value));
1037 }
1038 _ => bug!("found assignment info for non-assign statement"),
1039 }
1040 } else {
1041 self.super_statement(statement, location);
1042 }
1043 }
1044
1045 fn visit_operand(&mut self, operand: &mut Operand<'tcx>, location: Location) {
1046 match operand {
1047 Operand::Copy(place) | Operand::Move(place) => {
1048 if let Some(value) = self.before_effect.get(&(location, *place)) {
1049 *operand = self.make_operand(*value);
1050 } else if !place.projection.is_empty() {
1051 self.super_operand(operand, location)
1052 }
1053 }
1054 Operand::Constant(_) => {}
1055 }
1056 }
1057
1058 fn process_projection_elem(
1059 &mut self,
1060 elem: PlaceElem<'tcx>,
1061 location: Location,
1062 ) -> Option<PlaceElem<'tcx>> {
1063 if let PlaceElem::Index(local) = elem {
1064 let offset = self.before_effect.get(&(location, local.into()))?;
1065 let offset = offset.try_to_scalar()?;
1066 let offset = offset.to_target_usize(&self.tcx).discard_err()?;
1067 let min_length = offset.checked_add(1)?;
1068 Some(PlaceElem::ConstantIndex { offset, min_length, from_end: false })
1069 } else {
1070 None
1071 }
1072 }
1073}
1074
1075struct OperandCollector<'a, 'b, 'tcx> {
1076 state: &'a State<FlatSet<Scalar>>,
1077 visitor: &'a mut Collector<'b, 'tcx>,
1078 ecx: &'a mut InterpCx<'tcx, DummyMachine>,
1079 map: &'a Map<'tcx>,
1080}
1081
1082impl<'tcx> Visitor<'tcx> for OperandCollector<'_, '_, 'tcx> {
1083 fn visit_projection_elem(
1084 &mut self,
1085 _: PlaceRef<'tcx>,
1086 elem: PlaceElem<'tcx>,
1087 _: PlaceContext,
1088 location: Location,
1089 ) {
1090 if let PlaceElem::Index(local) = elem
1091 && let Some(value) =
1092 self.visitor.try_make_constant(self.ecx, local.into(), self.state, self.map)
1093 {
1094 self.visitor.patch.before_effect.insert((location, local.into()), value);
1095 }
1096 }
1097
1098 fn visit_operand(&mut self, operand: &Operand<'tcx>, location: Location) {
1099 if let Some(place) = operand.place() {
1100 if let Some(value) =
1101 self.visitor.try_make_constant(self.ecx, place, self.state, self.map)
1102 {
1103 self.visitor.patch.before_effect.insert((location, place), value);
1104 } else if !place.projection.is_empty() {
1105 self.super_operand(operand, location)
1107 }
1108 }
1109 }
1110}