use crate::{ fluent_generated as fluent, lints::{ AtomicOrderingFence, AtomicOrderingLoad, AtomicOrderingStore, ImproperCTypes, InvalidAtomicOrderingDiag, OnlyCastu8ToChar, OverflowingBinHex, OverflowingBinHexSign, OverflowingBinHexSub, OverflowingInt, OverflowingIntHelp, OverflowingLiteral, OverflowingUInt, RangeEndpointOutOfRange, UnusedComparisons, VariantSizeDifferencesDiag, }, }; use crate::{LateContext, LateLintPass, LintContext}; use rustc_ast as ast; use rustc_attr as attr; use rustc_data_structures::fx::FxHashSet; use rustc_errors::DiagnosticMessage; use rustc_hir as hir; use rustc_hir::{is_range_literal, Expr, ExprKind, Node}; use rustc_middle::ty::layout::{IntegerExt, LayoutOf, SizeSkeleton}; use rustc_middle::ty::subst::SubstsRef; use rustc_middle::ty::{ self, AdtKind, DefIdTree, Ty, TyCtxt, TypeSuperVisitable, TypeVisitable, TypeVisitableExt, }; use rustc_span::def_id::LocalDefId; use rustc_span::source_map; use rustc_span::symbol::sym; use rustc_span::{Span, Symbol}; use rustc_target::abi::{Abi, Size, WrappingRange}; use rustc_target::abi::{Integer, TagEncoding, Variants}; use rustc_target::spec::abi::Abi as SpecAbi; use std::iter; use std::ops::ControlFlow; declare_lint! { /// The `unused_comparisons` lint detects comparisons made useless by /// limits of the types involved. /// /// ### Example /// /// ```rust /// fn foo(x: u8) { /// x >= 0; /// } /// ``` /// /// {{produces}} /// /// ### Explanation /// /// A useless comparison may indicate a mistake, and should be fixed or /// removed. UNUSED_COMPARISONS, Warn, "comparisons made useless by limits of the types involved" } declare_lint! { /// The `overflowing_literals` lint detects literal out of range for its /// type. /// /// ### Example /// /// ```rust,compile_fail /// let x: u8 = 1000; /// ``` /// /// {{produces}} /// /// ### Explanation /// /// It is usually a mistake to use a literal that overflows the type where /// it is used. Either use a literal that is within range, or change the /// type to be within the range of the literal. OVERFLOWING_LITERALS, Deny, "literal out of range for its type" } declare_lint! { /// The `variant_size_differences` lint detects enums with widely varying /// variant sizes. /// /// ### Example /// /// ```rust,compile_fail /// #![deny(variant_size_differences)] /// enum En { /// V0(u8), /// VBig([u8; 1024]), /// } /// ``` /// /// {{produces}} /// /// ### Explanation /// /// It can be a mistake to add a variant to an enum that is much larger /// than the other variants, bloating the overall size required for all /// variants. This can impact performance and memory usage. This is /// triggered if one variant is more than 3 times larger than the /// second-largest variant. /// /// Consider placing the large variant's contents on the heap (for example /// via [`Box`]) to keep the overall size of the enum itself down. /// /// This lint is "allow" by default because it can be noisy, and may not be /// an actual problem. Decisions about this should be guided with /// profiling and benchmarking. /// /// [`Box`]: https://doc.rust-lang.org/std/boxed/index.html VARIANT_SIZE_DIFFERENCES, Allow, "detects enums with widely varying variant sizes" } #[derive(Copy, Clone)] pub struct TypeLimits { /// Id of the last visited negated expression negated_expr_id: Option, } impl_lint_pass!(TypeLimits => [UNUSED_COMPARISONS, OVERFLOWING_LITERALS]); impl TypeLimits { pub fn new() -> TypeLimits { TypeLimits { negated_expr_id: None } } } /// Attempts to special-case the overflowing literal lint when it occurs as a range endpoint (`expr..MAX+1`). /// Returns `true` iff the lint was emitted. fn lint_overflowing_range_endpoint<'tcx>( cx: &LateContext<'tcx>, lit: &hir::Lit, lit_val: u128, max: u128, expr: &'tcx hir::Expr<'tcx>, ty: &str, ) -> bool { // We only want to handle exclusive (`..`) ranges, // which are represented as `ExprKind::Struct`. let par_id = cx.tcx.hir().parent_id(expr.hir_id); let Node::ExprField(field) = cx.tcx.hir().get(par_id) else { return false }; let Node::Expr(struct_expr) = cx.tcx.hir().get_parent(field.hir_id) else { return false }; if !is_range_literal(struct_expr) { return false; }; let ExprKind::Struct(_, eps, _) = &struct_expr.kind else { return false }; if eps.len() != 2 { return false; } // We can suggest using an inclusive range // (`..=`) instead only if it is the `end` that is // overflowing and only by 1. if !(eps[1].expr.hir_id == expr.hir_id && lit_val - 1 == max) { return false; }; let Ok(start) = cx.sess().source_map().span_to_snippet(eps[0].span) else { return false }; use rustc_ast::{LitIntType, LitKind}; let suffix = match lit.node { LitKind::Int(_, LitIntType::Signed(s)) => s.name_str(), LitKind::Int(_, LitIntType::Unsigned(s)) => s.name_str(), LitKind::Int(_, LitIntType::Unsuffixed) => "", _ => bug!(), }; cx.emit_spanned_lint( OVERFLOWING_LITERALS, struct_expr.span, RangeEndpointOutOfRange { ty, suggestion: struct_expr.span, start, literal: lit_val - 1, suffix, }, ); // We've just emitted a lint, special cased for `(...)..MAX+1` ranges, // return `true` so the callers don't also emit a lint true } // For `isize` & `usize`, be conservative with the warnings, so that the // warnings are consistent between 32- and 64-bit platforms. fn int_ty_range(int_ty: ty::IntTy) -> (i128, i128) { match int_ty { ty::IntTy::Isize => (i64::MIN.into(), i64::MAX.into()), ty::IntTy::I8 => (i8::MIN.into(), i8::MAX.into()), ty::IntTy::I16 => (i16::MIN.into(), i16::MAX.into()), ty::IntTy::I32 => (i32::MIN.into(), i32::MAX.into()), ty::IntTy::I64 => (i64::MIN.into(), i64::MAX.into()), ty::IntTy::I128 => (i128::MIN, i128::MAX), } } fn uint_ty_range(uint_ty: ty::UintTy) -> (u128, u128) { let max = match uint_ty { ty::UintTy::Usize => u64::MAX.into(), ty::UintTy::U8 => u8::MAX.into(), ty::UintTy::U16 => u16::MAX.into(), ty::UintTy::U32 => u32::MAX.into(), ty::UintTy::U64 => u64::MAX.into(), ty::UintTy::U128 => u128::MAX, }; (0, max) } fn get_bin_hex_repr(cx: &LateContext<'_>, lit: &hir::Lit) -> Option { let src = cx.sess().source_map().span_to_snippet(lit.span).ok()?; let firstch = src.chars().next()?; if firstch == '0' { match src.chars().nth(1) { Some('x' | 'b') => return Some(src), _ => return None, } } None } fn report_bin_hex_error( cx: &LateContext<'_>, expr: &hir::Expr<'_>, ty: attr::IntType, size: Size, repr_str: String, val: u128, negative: bool, ) { let (t, actually) = match ty { attr::IntType::SignedInt(t) => { let actually = if negative { -(size.sign_extend(val) as i128) } else { size.sign_extend(val) as i128 }; (t.name_str(), actually.to_string()) } attr::IntType::UnsignedInt(t) => { let actually = size.truncate(val); (t.name_str(), actually.to_string()) } }; let sign = if negative { OverflowingBinHexSign::Negative } else { OverflowingBinHexSign::Positive }; let sub = get_type_suggestion(cx.typeck_results().node_type(expr.hir_id), val, negative).map( |suggestion_ty| { if let Some(pos) = repr_str.chars().position(|c| c == 'i' || c == 'u') { let (sans_suffix, _) = repr_str.split_at(pos); OverflowingBinHexSub::Suggestion { span: expr.span, suggestion_ty, sans_suffix } } else { OverflowingBinHexSub::Help { suggestion_ty } } }, ); cx.emit_spanned_lint( OVERFLOWING_LITERALS, expr.span, OverflowingBinHex { ty: t, lit: repr_str.clone(), dec: val, actually, sign, sub }, ) } // This function finds the next fitting type and generates a suggestion string. // It searches for fitting types in the following way (`X < Y`): // - `iX`: if literal fits in `uX` => `uX`, else => `iY` // - `-iX` => `iY` // - `uX` => `uY` // // No suggestion for: `isize`, `usize`. fn get_type_suggestion(t: Ty<'_>, val: u128, negative: bool) -> Option<&'static str> { use ty::IntTy::*; use ty::UintTy::*; macro_rules! find_fit { ($ty:expr, $val:expr, $negative:expr, $($type:ident => [$($utypes:expr),*] => [$($itypes:expr),*]),+) => { { let _neg = if negative { 1 } else { 0 }; match $ty { $($type => { $(if !negative && val <= uint_ty_range($utypes).1 { return Some($utypes.name_str()) })* $(if val <= int_ty_range($itypes).1 as u128 + _neg { return Some($itypes.name_str()) })* None },)+ _ => None } } } } match t.kind() { ty::Int(i) => find_fit!(i, val, negative, I8 => [U8] => [I16, I32, I64, I128], I16 => [U16] => [I32, I64, I128], I32 => [U32] => [I64, I128], I64 => [U64] => [I128], I128 => [U128] => []), ty::Uint(u) => find_fit!(u, val, negative, U8 => [U8, U16, U32, U64, U128] => [], U16 => [U16, U32, U64, U128] => [], U32 => [U32, U64, U128] => [], U64 => [U64, U128] => [], U128 => [U128] => []), _ => None, } } fn lint_int_literal<'tcx>( cx: &LateContext<'tcx>, type_limits: &TypeLimits, e: &'tcx hir::Expr<'tcx>, lit: &hir::Lit, t: ty::IntTy, v: u128, ) { let int_type = t.normalize(cx.sess().target.pointer_width); let (min, max) = int_ty_range(int_type); let max = max as u128; let negative = type_limits.negated_expr_id == Some(e.hir_id); // Detect literal value out of range [min, max] inclusive // avoiding use of -min to prevent overflow/panic if (negative && v > max + 1) || (!negative && v > max) { if let Some(repr_str) = get_bin_hex_repr(cx, lit) { report_bin_hex_error( cx, e, attr::IntType::SignedInt(ty::ast_int_ty(t)), Integer::from_int_ty(cx, t).size(), repr_str, v, negative, ); return; } if lint_overflowing_range_endpoint(cx, lit, v, max, e, t.name_str()) { // The overflowing literal lint was emitted by `lint_overflowing_range_endpoint`. return; } let lit = cx .sess() .source_map() .span_to_snippet(lit.span) .expect("must get snippet from literal"); let help = get_type_suggestion(cx.typeck_results().node_type(e.hir_id), v, negative) .map(|suggestion_ty| OverflowingIntHelp { suggestion_ty }); cx.emit_spanned_lint( OVERFLOWING_LITERALS, e.span, OverflowingInt { ty: t.name_str(), lit, min, max, help }, ); } } fn lint_uint_literal<'tcx>( cx: &LateContext<'tcx>, e: &'tcx hir::Expr<'tcx>, lit: &hir::Lit, t: ty::UintTy, ) { let uint_type = t.normalize(cx.sess().target.pointer_width); let (min, max) = uint_ty_range(uint_type); let lit_val: u128 = match lit.node { // _v is u8, within range by definition ast::LitKind::Byte(_v) => return, ast::LitKind::Int(v, _) => v, _ => bug!(), }; if lit_val < min || lit_val > max { let parent_id = cx.tcx.hir().parent_id(e.hir_id); if let Node::Expr(par_e) = cx.tcx.hir().get(parent_id) { match par_e.kind { hir::ExprKind::Cast(..) => { if let ty::Char = cx.typeck_results().expr_ty(par_e).kind() { cx.emit_spanned_lint( OVERFLOWING_LITERALS, par_e.span, OnlyCastu8ToChar { span: par_e.span, literal: lit_val }, ); return; } } _ => {} } } if lint_overflowing_range_endpoint(cx, lit, lit_val, max, e, t.name_str()) { // The overflowing literal lint was emitted by `lint_overflowing_range_endpoint`. return; } if let Some(repr_str) = get_bin_hex_repr(cx, lit) { report_bin_hex_error( cx, e, attr::IntType::UnsignedInt(ty::ast_uint_ty(t)), Integer::from_uint_ty(cx, t).size(), repr_str, lit_val, false, ); return; } cx.emit_spanned_lint( OVERFLOWING_LITERALS, e.span, OverflowingUInt { ty: t.name_str(), lit: cx .sess() .source_map() .span_to_snippet(lit.span) .expect("must get snippet from literal"), min, max, }, ); } } fn lint_literal<'tcx>( cx: &LateContext<'tcx>, type_limits: &TypeLimits, e: &'tcx hir::Expr<'tcx>, lit: &hir::Lit, ) { match *cx.typeck_results().node_type(e.hir_id).kind() { ty::Int(t) => { match lit.node { ast::LitKind::Int(v, ast::LitIntType::Signed(_) | ast::LitIntType::Unsuffixed) => { lint_int_literal(cx, type_limits, e, lit, t, v) } _ => bug!(), }; } ty::Uint(t) => lint_uint_literal(cx, e, lit, t), ty::Float(t) => { let is_infinite = match lit.node { ast::LitKind::Float(v, _) => match t { ty::FloatTy::F32 => v.as_str().parse().map(f32::is_infinite), ty::FloatTy::F64 => v.as_str().parse().map(f64::is_infinite), }, _ => bug!(), }; if is_infinite == Ok(true) { cx.emit_spanned_lint( OVERFLOWING_LITERALS, e.span, OverflowingLiteral { ty: t.name_str(), lit: cx .sess() .source_map() .span_to_snippet(lit.span) .expect("must get snippet from literal"), }, ); } } _ => {} } } impl<'tcx> LateLintPass<'tcx> for TypeLimits { fn check_expr(&mut self, cx: &LateContext<'tcx>, e: &'tcx hir::Expr<'tcx>) { match e.kind { hir::ExprKind::Unary(hir::UnOp::Neg, ref expr) => { // propagate negation, if the negation itself isn't negated if self.negated_expr_id != Some(e.hir_id) { self.negated_expr_id = Some(expr.hir_id); } } hir::ExprKind::Binary(binop, ref l, ref r) => { if is_comparison(binop) && !check_limits(cx, binop, &l, &r) { cx.emit_spanned_lint(UNUSED_COMPARISONS, e.span, UnusedComparisons); } } hir::ExprKind::Lit(ref lit) => lint_literal(cx, self, e, lit), _ => {} }; fn is_valid(binop: hir::BinOp, v: T, min: T, max: T) -> bool { match binop.node { hir::BinOpKind::Lt => v > min && v <= max, hir::BinOpKind::Le => v >= min && v < max, hir::BinOpKind::Gt => v >= min && v < max, hir::BinOpKind::Ge => v > min && v <= max, hir::BinOpKind::Eq | hir::BinOpKind::Ne => v >= min && v <= max, _ => bug!(), } } fn rev_binop(binop: hir::BinOp) -> hir::BinOp { source_map::respan( binop.span, match binop.node { hir::BinOpKind::Lt => hir::BinOpKind::Gt, hir::BinOpKind::Le => hir::BinOpKind::Ge, hir::BinOpKind::Gt => hir::BinOpKind::Lt, hir::BinOpKind::Ge => hir::BinOpKind::Le, _ => return binop, }, ) } fn check_limits( cx: &LateContext<'_>, binop: hir::BinOp, l: &hir::Expr<'_>, r: &hir::Expr<'_>, ) -> bool { let (lit, expr, swap) = match (&l.kind, &r.kind) { (&hir::ExprKind::Lit(_), _) => (l, r, true), (_, &hir::ExprKind::Lit(_)) => (r, l, false), _ => return true, }; // Normalize the binop so that the literal is always on the RHS in // the comparison let norm_binop = if swap { rev_binop(binop) } else { binop }; match *cx.typeck_results().node_type(expr.hir_id).kind() { ty::Int(int_ty) => { let (min, max) = int_ty_range(int_ty); let lit_val: i128 = match lit.kind { hir::ExprKind::Lit(ref li) => match li.node { ast::LitKind::Int( v, ast::LitIntType::Signed(_) | ast::LitIntType::Unsuffixed, ) => v as i128, _ => return true, }, _ => bug!(), }; is_valid(norm_binop, lit_val, min, max) } ty::Uint(uint_ty) => { let (min, max): (u128, u128) = uint_ty_range(uint_ty); let lit_val: u128 = match lit.kind { hir::ExprKind::Lit(ref li) => match li.node { ast::LitKind::Int(v, _) => v, _ => return true, }, _ => bug!(), }; is_valid(norm_binop, lit_val, min, max) } _ => true, } } fn is_comparison(binop: hir::BinOp) -> bool { matches!( binop.node, hir::BinOpKind::Eq | hir::BinOpKind::Lt | hir::BinOpKind::Le | hir::BinOpKind::Ne | hir::BinOpKind::Ge | hir::BinOpKind::Gt ) } } } declare_lint! { /// The `improper_ctypes` lint detects incorrect use of types in foreign /// modules. /// /// ### Example /// /// ```rust /// extern "C" { /// static STATIC: String; /// } /// ``` /// /// {{produces}} /// /// ### Explanation /// /// The compiler has several checks to verify that types used in `extern` /// blocks are safe and follow certain rules to ensure proper /// compatibility with the foreign interfaces. This lint is issued when it /// detects a probable mistake in a definition. The lint usually should /// provide a description of the issue, along with possibly a hint on how /// to resolve it. IMPROPER_CTYPES, Warn, "proper use of libc types in foreign modules" } declare_lint_pass!(ImproperCTypesDeclarations => [IMPROPER_CTYPES]); declare_lint! { /// The `improper_ctypes_definitions` lint detects incorrect use of /// [`extern` function] definitions. /// /// [`extern` function]: https://doc.rust-lang.org/reference/items/functions.html#extern-function-qualifier /// /// ### Example /// /// ```rust /// # #![allow(unused)] /// pub extern "C" fn str_type(p: &str) { } /// ``` /// /// {{produces}} /// /// ### Explanation /// /// There are many parameter and return types that may be specified in an /// `extern` function that are not compatible with the given ABI. This /// lint is an alert that these types should not be used. The lint usually /// should provide a description of the issue, along with possibly a hint /// on how to resolve it. IMPROPER_CTYPES_DEFINITIONS, Warn, "proper use of libc types in foreign item definitions" } declare_lint_pass!(ImproperCTypesDefinitions => [IMPROPER_CTYPES_DEFINITIONS]); #[derive(Clone, Copy)] pub(crate) enum CItemKind { Declaration, Definition, } struct ImproperCTypesVisitor<'a, 'tcx> { cx: &'a LateContext<'tcx>, mode: CItemKind, } enum FfiResult<'tcx> { FfiSafe, FfiPhantom(Ty<'tcx>), FfiUnsafe { ty: Ty<'tcx>, reason: DiagnosticMessage, help: Option }, } pub(crate) fn nonnull_optimization_guaranteed<'tcx>( tcx: TyCtxt<'tcx>, def: ty::AdtDef<'tcx>, ) -> bool { tcx.has_attr(def.did(), sym::rustc_nonnull_optimization_guaranteed) } /// `repr(transparent)` structs can have a single non-ZST field, this function returns that /// field. pub fn transparent_newtype_field<'a, 'tcx>( tcx: TyCtxt<'tcx>, variant: &'a ty::VariantDef, ) -> Option<&'a ty::FieldDef> { let param_env = tcx.param_env(variant.def_id); variant.fields.iter().find(|field| { let field_ty = tcx.type_of(field.did).subst_identity(); let is_zst = tcx.layout_of(param_env.and(field_ty)).map_or(false, |layout| layout.is_zst()); !is_zst }) } /// Is type known to be non-null? fn ty_is_known_nonnull<'tcx>(cx: &LateContext<'tcx>, ty: Ty<'tcx>, mode: CItemKind) -> bool { let tcx = cx.tcx; match ty.kind() { ty::FnPtr(_) => true, ty::Ref(..) => true, ty::Adt(def, _) if def.is_box() && matches!(mode, CItemKind::Definition) => true, ty::Adt(def, substs) if def.repr().transparent() && !def.is_union() => { let marked_non_null = nonnull_optimization_guaranteed(tcx, *def); if marked_non_null { return true; } // `UnsafeCell` has its niche hidden. if def.is_unsafe_cell() { return false; } def.variants() .iter() .filter_map(|variant| transparent_newtype_field(cx.tcx, variant)) .any(|field| ty_is_known_nonnull(cx, field.ty(tcx, substs), mode)) } _ => false, } } /// Given a non-null scalar (or transparent) type `ty`, return the nullable version of that type. /// If the type passed in was not scalar, returns None. fn get_nullable_type<'tcx>(cx: &LateContext<'tcx>, ty: Ty<'tcx>) -> Option> { let tcx = cx.tcx; Some(match *ty.kind() { ty::Adt(field_def, field_substs) => { let inner_field_ty = { let mut first_non_zst_ty = field_def .variants() .iter() .filter_map(|v| transparent_newtype_field(cx.tcx, v)); debug_assert_eq!( first_non_zst_ty.clone().count(), 1, "Wrong number of fields for transparent type" ); first_non_zst_ty .next_back() .expect("No non-zst fields in transparent type.") .ty(tcx, field_substs) }; return get_nullable_type(cx, inner_field_ty); } ty::Int(ty) => tcx.mk_mach_int(ty), ty::Uint(ty) => tcx.mk_mach_uint(ty), ty::RawPtr(ty_mut) => tcx.mk_ptr(ty_mut), // As these types are always non-null, the nullable equivalent of // Option of these types are their raw pointer counterparts. ty::Ref(_region, ty, mutbl) => tcx.mk_ptr(ty::TypeAndMut { ty, mutbl }), ty::FnPtr(..) => { // There is no nullable equivalent for Rust's function pointers -- you // must use an Option _> to represent it. ty } // We should only ever reach this case if ty_is_known_nonnull is extended // to other types. ref unhandled => { debug!( "get_nullable_type: Unhandled scalar kind: {:?} while checking {:?}", unhandled, ty ); return None; } }) } /// Check if this enum can be safely exported based on the "nullable pointer optimization". If it /// can, return the type that `ty` can be safely converted to, otherwise return `None`. /// Currently restricted to function pointers, boxes, references, `core::num::NonZero*`, /// `core::ptr::NonNull`, and `#[repr(transparent)]` newtypes. /// FIXME: This duplicates code in codegen. pub(crate) fn repr_nullable_ptr<'tcx>( cx: &LateContext<'tcx>, ty: Ty<'tcx>, ckind: CItemKind, ) -> Option> { debug!("is_repr_nullable_ptr(cx, ty = {:?})", ty); if let ty::Adt(ty_def, substs) = ty.kind() { let field_ty = match &ty_def.variants().raw[..] { [var_one, var_two] => match (&var_one.fields[..], &var_two.fields[..]) { ([], [field]) | ([field], []) => field.ty(cx.tcx, substs), _ => return None, }, _ => return None, }; if !ty_is_known_nonnull(cx, field_ty, ckind) { return None; } // At this point, the field's type is known to be nonnull and the parent enum is Option-like. // If the computed size for the field and the enum are different, the nonnull optimization isn't // being applied (and we've got a problem somewhere). let compute_size_skeleton = |t| SizeSkeleton::compute(t, cx.tcx, cx.param_env).unwrap(); if !compute_size_skeleton(ty).same_size(compute_size_skeleton(field_ty)) { bug!("improper_ctypes: Option nonnull optimization not applied?"); } // Return the nullable type this Option-like enum can be safely represented with. let field_ty_abi = &cx.layout_of(field_ty).unwrap().abi; if let Abi::Scalar(field_ty_scalar) = field_ty_abi { match field_ty_scalar.valid_range(cx) { WrappingRange { start: 0, end } if end == field_ty_scalar.size(&cx.tcx).unsigned_int_max() - 1 => { return Some(get_nullable_type(cx, field_ty).unwrap()); } WrappingRange { start: 1, .. } => { return Some(get_nullable_type(cx, field_ty).unwrap()); } WrappingRange { start, end } => { unreachable!("Unhandled start and end range: ({}, {})", start, end) } }; } } None } impl<'a, 'tcx> ImproperCTypesVisitor<'a, 'tcx> { /// Check if the type is array and emit an unsafe type lint. fn check_for_array_ty(&mut self, sp: Span, ty: Ty<'tcx>) -> bool { if let ty::Array(..) = ty.kind() { self.emit_ffi_unsafe_type_lint( ty, sp, fluent::lint_improper_ctypes_array_reason, Some(fluent::lint_improper_ctypes_array_help), ); true } else { false } } /// Checks if the given field's type is "ffi-safe". fn check_field_type_for_ffi( &self, cache: &mut FxHashSet>, field: &ty::FieldDef, substs: SubstsRef<'tcx>, ) -> FfiResult<'tcx> { let field_ty = field.ty(self.cx.tcx, substs); if field_ty.has_opaque_types() { self.check_type_for_ffi(cache, field_ty) } else { let field_ty = self.cx.tcx.normalize_erasing_regions(self.cx.param_env, field_ty); self.check_type_for_ffi(cache, field_ty) } } /// Checks if the given `VariantDef`'s field types are "ffi-safe". fn check_variant_for_ffi( &self, cache: &mut FxHashSet>, ty: Ty<'tcx>, def: ty::AdtDef<'tcx>, variant: &ty::VariantDef, substs: SubstsRef<'tcx>, ) -> FfiResult<'tcx> { use FfiResult::*; let transparent_safety = def.repr().transparent().then(|| { // Can assume that at most one field is not a ZST, so only check // that field's type for FFI-safety. if let Some(field) = transparent_newtype_field(self.cx.tcx, variant) { return self.check_field_type_for_ffi(cache, field, substs); } else { // All fields are ZSTs; this means that the type should behave // like (), which is FFI-unsafe... except if all fields are PhantomData, // which is tested for below FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_struct_zst, help: None } } }); // We can't completely trust repr(C) markings; make sure the fields are // actually safe. let mut all_phantom = !variant.fields.is_empty(); for field in &variant.fields { match self.check_field_type_for_ffi(cache, &field, substs) { FfiSafe => { all_phantom = false; } FfiPhantom(..) if !def.repr().transparent() && def.is_enum() => { return FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_enum_phantomdata, help: None, }; } FfiPhantom(..) => {} r => return transparent_safety.unwrap_or(r), } } if all_phantom { FfiPhantom(ty) } else { transparent_safety.unwrap_or(FfiSafe) } } /// Checks if the given type is "ffi-safe" (has a stable, well-defined /// representation which can be exported to C code). fn check_type_for_ffi(&self, cache: &mut FxHashSet>, ty: Ty<'tcx>) -> FfiResult<'tcx> { use FfiResult::*; let tcx = self.cx.tcx; // Protect against infinite recursion, for example // `struct S(*mut S);`. // FIXME: A recursion limit is necessary as well, for irregular // recursive types. if !cache.insert(ty) { return FfiSafe; } match *ty.kind() { ty::Adt(def, substs) => { if def.is_box() && matches!(self.mode, CItemKind::Definition) { if ty.boxed_ty().is_sized(tcx, self.cx.param_env) { return FfiSafe; } else { return FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_box, help: None, }; } } if def.is_phantom_data() { return FfiPhantom(ty); } match def.adt_kind() { AdtKind::Struct | AdtKind::Union => { if !def.repr().c() && !def.repr().transparent() { return FfiUnsafe { ty, reason: if def.is_struct() { fluent::lint_improper_ctypes_struct_layout_reason } else { fluent::lint_improper_ctypes_union_layout_reason }, help: if def.is_struct() { Some(fluent::lint_improper_ctypes_struct_layout_help) } else { Some(fluent::lint_improper_ctypes_union_layout_help) }, }; } let is_non_exhaustive = def.non_enum_variant().is_field_list_non_exhaustive(); if is_non_exhaustive && !def.did().is_local() { return FfiUnsafe { ty, reason: if def.is_struct() { fluent::lint_improper_ctypes_struct_non_exhaustive } else { fluent::lint_improper_ctypes_union_non_exhaustive }, help: None, }; } if def.non_enum_variant().fields.is_empty() { return FfiUnsafe { ty, reason: if def.is_struct() { fluent::lint_improper_ctypes_struct_fieldless_reason } else { fluent::lint_improper_ctypes_union_fieldless_reason }, help: if def.is_struct() { Some(fluent::lint_improper_ctypes_struct_fieldless_help) } else { Some(fluent::lint_improper_ctypes_union_fieldless_help) }, }; } self.check_variant_for_ffi(cache, ty, def, def.non_enum_variant(), substs) } AdtKind::Enum => { if def.variants().is_empty() { // Empty enums are okay... although sort of useless. return FfiSafe; } // Check for a repr() attribute to specify the size of the // discriminant. if !def.repr().c() && !def.repr().transparent() && def.repr().int.is_none() { // Special-case types like `Option`. if repr_nullable_ptr(self.cx, ty, self.mode).is_none() { return FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_enum_repr_reason, help: Some(fluent::lint_improper_ctypes_enum_repr_help), }; } } if def.is_variant_list_non_exhaustive() && !def.did().is_local() { return FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_non_exhaustive, help: None, }; } // Check the contained variants. for variant in def.variants() { let is_non_exhaustive = variant.is_field_list_non_exhaustive(); if is_non_exhaustive && !variant.def_id.is_local() { return FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_non_exhaustive_variant, help: None, }; } match self.check_variant_for_ffi(cache, ty, def, variant, substs) { FfiSafe => (), r => return r, } } FfiSafe } } } ty::Char => FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_char_reason, help: Some(fluent::lint_improper_ctypes_char_help), }, ty::Int(ty::IntTy::I128) | ty::Uint(ty::UintTy::U128) => { FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_128bit, help: None } } // Primitive types with a stable representation. ty::Bool | ty::Int(..) | ty::Uint(..) | ty::Float(..) | ty::Never => FfiSafe, ty::Slice(_) => FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_slice_reason, help: Some(fluent::lint_improper_ctypes_slice_help), }, ty::Dynamic(..) => { FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_dyn, help: None } } ty::Str => FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_str_reason, help: Some(fluent::lint_improper_ctypes_str_help), }, ty::Tuple(..) => FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_tuple_reason, help: Some(fluent::lint_improper_ctypes_tuple_help), }, ty::RawPtr(ty::TypeAndMut { ty, .. }) | ty::Ref(_, ty, _) if { matches!(self.mode, CItemKind::Definition) && ty.is_sized(self.cx.tcx, self.cx.param_env) } => { FfiSafe } ty::RawPtr(ty::TypeAndMut { ty, .. }) if match ty.kind() { ty::Tuple(tuple) => tuple.is_empty(), _ => false, } => { FfiSafe } ty::RawPtr(ty::TypeAndMut { ty, .. }) | ty::Ref(_, ty, _) => { self.check_type_for_ffi(cache, ty) } ty::Array(inner_ty, _) => self.check_type_for_ffi(cache, inner_ty), ty::FnPtr(sig) => { if self.is_internal_abi(sig.abi()) { return FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_fnptr_reason, help: Some(fluent::lint_improper_ctypes_fnptr_help), }; } let sig = tcx.erase_late_bound_regions(sig); if !sig.output().is_unit() { let r = self.check_type_for_ffi(cache, sig.output()); match r { FfiSafe => {} _ => { return r; } } } for arg in sig.inputs() { let r = self.check_type_for_ffi(cache, *arg); match r { FfiSafe => {} _ => { return r; } } } FfiSafe } ty::Foreign(..) => FfiSafe, // While opaque types are checked for earlier, if a projection in a struct field // normalizes to an opaque type, then it will reach this branch. ty::Alias(ty::Opaque, ..) => { FfiUnsafe { ty, reason: fluent::lint_improper_ctypes_opaque, help: None } } // `extern "C" fn` functions can have type parameters, which may or may not be FFI-safe, // so they are currently ignored for the purposes of this lint. ty::Param(..) | ty::Alias(ty::Projection, ..) if matches!(self.mode, CItemKind::Definition) => { FfiSafe } ty::Param(..) | ty::Alias(ty::Projection, ..) | ty::Infer(..) | ty::Bound(..) | ty::Error(_) | ty::Closure(..) | ty::Generator(..) | ty::GeneratorWitness(..) | ty::GeneratorWitnessMIR(..) | ty::Placeholder(..) | ty::FnDef(..) => bug!("unexpected type in foreign function: {:?}", ty), } } fn emit_ffi_unsafe_type_lint( &mut self, ty: Ty<'tcx>, sp: Span, note: DiagnosticMessage, help: Option, ) { let lint = match self.mode { CItemKind::Declaration => IMPROPER_CTYPES, CItemKind::Definition => IMPROPER_CTYPES_DEFINITIONS, }; let desc = match self.mode { CItemKind::Declaration => "block", CItemKind::Definition => "fn", }; let span_note = if let ty::Adt(def, _) = ty.kind() && let Some(sp) = self.cx.tcx.hir().span_if_local(def.did()) { Some(sp) } else { None }; self.cx.emit_spanned_lint( lint, sp, ImproperCTypes { ty, desc, label: sp, help, note, span_note }, ); } fn check_for_opaque_ty(&mut self, sp: Span, ty: Ty<'tcx>) -> bool { struct ProhibitOpaqueTypes; impl<'tcx> ty::visit::TypeVisitor> for ProhibitOpaqueTypes { type BreakTy = Ty<'tcx>; fn visit_ty(&mut self, ty: Ty<'tcx>) -> ControlFlow { if !ty.has_opaque_types() { return ControlFlow::Continue(()); } if let ty::Alias(ty::Opaque, ..) = ty.kind() { ControlFlow::Break(ty) } else { ty.super_visit_with(self) } } } if let Some(ty) = self .cx .tcx .normalize_erasing_regions(self.cx.param_env, ty) .visit_with(&mut ProhibitOpaqueTypes) .break_value() { self.emit_ffi_unsafe_type_lint(ty, sp, fluent::lint_improper_ctypes_opaque, None); true } else { false } } fn check_type_for_ffi_and_report_errors( &mut self, sp: Span, ty: Ty<'tcx>, is_static: bool, is_return_type: bool, ) { // We have to check for opaque types before `normalize_erasing_regions`, // which will replace opaque types with their underlying concrete type. if self.check_for_opaque_ty(sp, ty) { // We've already emitted an error due to an opaque type. return; } // it is only OK to use this function because extern fns cannot have // any generic types right now: let ty = self.cx.tcx.normalize_erasing_regions(self.cx.param_env, ty); // C doesn't really support passing arrays by value - the only way to pass an array by value // is through a struct. So, first test that the top level isn't an array, and then // recursively check the types inside. if !is_static && self.check_for_array_ty(sp, ty) { return; } // Don't report FFI errors for unit return types. This check exists here, and not in // `check_foreign_fn` (where it would make more sense) so that normalization has definitely // happened. if is_return_type && ty.is_unit() { return; } match self.check_type_for_ffi(&mut FxHashSet::default(), ty) { FfiResult::FfiSafe => {} FfiResult::FfiPhantom(ty) => { self.emit_ffi_unsafe_type_lint( ty, sp, fluent::lint_improper_ctypes_only_phantomdata, None, ); } // If `ty` is a `repr(transparent)` newtype, and the non-zero-sized type is a generic // argument, which after substitution, is `()`, then this branch can be hit. FfiResult::FfiUnsafe { ty, .. } if is_return_type && ty.is_unit() => {} FfiResult::FfiUnsafe { ty, reason, help } => { self.emit_ffi_unsafe_type_lint(ty, sp, reason, help); } } } fn check_foreign_fn(&mut self, def_id: LocalDefId, decl: &hir::FnDecl<'_>) { let sig = self.cx.tcx.fn_sig(def_id).subst_identity(); let sig = self.cx.tcx.erase_late_bound_regions(sig); for (input_ty, input_hir) in iter::zip(sig.inputs(), decl.inputs) { self.check_type_for_ffi_and_report_errors(input_hir.span, *input_ty, false, false); } if let hir::FnRetTy::Return(ref ret_hir) = decl.output { let ret_ty = sig.output(); self.check_type_for_ffi_and_report_errors(ret_hir.span, ret_ty, false, true); } } fn check_foreign_static(&mut self, id: hir::OwnerId, span: Span) { let ty = self.cx.tcx.type_of(id).subst_identity(); self.check_type_for_ffi_and_report_errors(span, ty, true, false); } fn is_internal_abi(&self, abi: SpecAbi) -> bool { matches!( abi, SpecAbi::Rust | SpecAbi::RustCall | SpecAbi::RustIntrinsic | SpecAbi::PlatformIntrinsic ) } } impl<'tcx> LateLintPass<'tcx> for ImproperCTypesDeclarations { fn check_foreign_item(&mut self, cx: &LateContext<'_>, it: &hir::ForeignItem<'_>) { let mut vis = ImproperCTypesVisitor { cx, mode: CItemKind::Declaration }; let abi = cx.tcx.hir().get_foreign_abi(it.hir_id()); if !vis.is_internal_abi(abi) { match it.kind { hir::ForeignItemKind::Fn(ref decl, _, _) => { vis.check_foreign_fn(it.owner_id.def_id, decl); } hir::ForeignItemKind::Static(ref ty, _) => { vis.check_foreign_static(it.owner_id, ty.span); } hir::ForeignItemKind::Type => (), } } } } impl<'tcx> LateLintPass<'tcx> for ImproperCTypesDefinitions { fn check_fn( &mut self, cx: &LateContext<'tcx>, kind: hir::intravisit::FnKind<'tcx>, decl: &'tcx hir::FnDecl<'_>, _: &'tcx hir::Body<'_>, _: Span, id: LocalDefId, ) { use hir::intravisit::FnKind; let abi = match kind { FnKind::ItemFn(_, _, header, ..) => header.abi, FnKind::Method(_, sig, ..) => sig.header.abi, _ => return, }; let mut vis = ImproperCTypesVisitor { cx, mode: CItemKind::Definition }; if !vis.is_internal_abi(abi) { vis.check_foreign_fn(id, decl); } } } declare_lint_pass!(VariantSizeDifferences => [VARIANT_SIZE_DIFFERENCES]); impl<'tcx> LateLintPass<'tcx> for VariantSizeDifferences { fn check_item(&mut self, cx: &LateContext<'_>, it: &hir::Item<'_>) { if let hir::ItemKind::Enum(ref enum_definition, _) = it.kind { let t = cx.tcx.type_of(it.owner_id).subst_identity(); let ty = cx.tcx.erase_regions(t); let Ok(layout) = cx.layout_of(ty) else { return }; let Variants::Multiple { tag_encoding: TagEncoding::Direct, tag, ref variants, .. } = &layout.variants else { return }; let tag_size = tag.size(&cx.tcx).bytes(); debug!( "enum `{}` is {} bytes large with layout:\n{:#?}", t, layout.size.bytes(), layout ); let (largest, slargest, largest_index) = iter::zip(enum_definition.variants, variants) .map(|(variant, variant_layout)| { // Subtract the size of the enum tag. let bytes = variant_layout.size.bytes().saturating_sub(tag_size); debug!("- variant `{}` is {} bytes large", variant.ident, bytes); bytes }) .enumerate() .fold((0, 0, 0), |(l, s, li), (idx, size)| { if size > l { (size, l, idx) } else if size > s { (l, size, li) } else { (l, s, li) } }); // We only warn if the largest variant is at least thrice as large as // the second-largest. if largest > slargest * 3 && slargest > 0 { cx.emit_spanned_lint( VARIANT_SIZE_DIFFERENCES, enum_definition.variants[largest_index].span, VariantSizeDifferencesDiag { largest }, ); } } } } declare_lint! { /// The `invalid_atomic_ordering` lint detects passing an `Ordering` /// to an atomic operation that does not support that ordering. /// /// ### Example /// /// ```rust,compile_fail /// # use core::sync::atomic::{AtomicU8, Ordering}; /// let atom = AtomicU8::new(0); /// let value = atom.load(Ordering::Release); /// # let _ = value; /// ``` /// /// {{produces}} /// /// ### Explanation /// /// Some atomic operations are only supported for a subset of the /// `atomic::Ordering` variants. Passing an unsupported variant will cause /// an unconditional panic at runtime, which is detected by this lint. /// /// This lint will trigger in the following cases: (where `AtomicType` is an /// atomic type from `core::sync::atomic`, such as `AtomicBool`, /// `AtomicPtr`, `AtomicUsize`, or any of the other integer atomics). /// /// - Passing `Ordering::Acquire` or `Ordering::AcqRel` to /// `AtomicType::store`. /// /// - Passing `Ordering::Release` or `Ordering::AcqRel` to /// `AtomicType::load`. /// /// - Passing `Ordering::Relaxed` to `core::sync::atomic::fence` or /// `core::sync::atomic::compiler_fence`. /// /// - Passing `Ordering::Release` or `Ordering::AcqRel` as the failure /// ordering for any of `AtomicType::compare_exchange`, /// `AtomicType::compare_exchange_weak`, or `AtomicType::fetch_update`. INVALID_ATOMIC_ORDERING, Deny, "usage of invalid atomic ordering in atomic operations and memory fences" } declare_lint_pass!(InvalidAtomicOrdering => [INVALID_ATOMIC_ORDERING]); impl InvalidAtomicOrdering { fn inherent_atomic_method_call<'hir>( cx: &LateContext<'_>, expr: &Expr<'hir>, recognized_names: &[Symbol], // used for fast path calculation ) -> Option<(Symbol, &'hir [Expr<'hir>])> { const ATOMIC_TYPES: &[Symbol] = &[ sym::AtomicBool, sym::AtomicPtr, sym::AtomicUsize, sym::AtomicU8, sym::AtomicU16, sym::AtomicU32, sym::AtomicU64, sym::AtomicU128, sym::AtomicIsize, sym::AtomicI8, sym::AtomicI16, sym::AtomicI32, sym::AtomicI64, sym::AtomicI128, ]; if let ExprKind::MethodCall(ref method_path, _, args, _) = &expr.kind && recognized_names.contains(&method_path.ident.name) && let Some(m_def_id) = cx.typeck_results().type_dependent_def_id(expr.hir_id) && let Some(impl_did) = cx.tcx.impl_of_method(m_def_id) && let Some(adt) = cx.tcx.type_of(impl_did).subst_identity().ty_adt_def() // skip extension traits, only lint functions from the standard library && cx.tcx.trait_id_of_impl(impl_did).is_none() && let parent = cx.tcx.parent(adt.did()) && cx.tcx.is_diagnostic_item(sym::atomic_mod, parent) && ATOMIC_TYPES.contains(&cx.tcx.item_name(adt.did())) { return Some((method_path.ident.name, args)); } None } fn match_ordering(cx: &LateContext<'_>, ord_arg: &Expr<'_>) -> Option { let ExprKind::Path(ref ord_qpath) = ord_arg.kind else { return None }; let did = cx.qpath_res(ord_qpath, ord_arg.hir_id).opt_def_id()?; let tcx = cx.tcx; let atomic_ordering = tcx.get_diagnostic_item(sym::Ordering); let name = tcx.item_name(did); let parent = tcx.parent(did); [sym::Relaxed, sym::Release, sym::Acquire, sym::AcqRel, sym::SeqCst].into_iter().find( |&ordering| { name == ordering && (Some(parent) == atomic_ordering // needed in case this is a ctor, not a variant || tcx.opt_parent(parent) == atomic_ordering) }, ) } fn check_atomic_load_store(cx: &LateContext<'_>, expr: &Expr<'_>) { if let Some((method, args)) = Self::inherent_atomic_method_call(cx, expr, &[sym::load, sym::store]) && let Some((ordering_arg, invalid_ordering)) = match method { sym::load => Some((&args[0], sym::Release)), sym::store => Some((&args[1], sym::Acquire)), _ => None, } && let Some(ordering) = Self::match_ordering(cx, ordering_arg) && (ordering == invalid_ordering || ordering == sym::AcqRel) { if method == sym::load { cx.emit_spanned_lint(INVALID_ATOMIC_ORDERING, ordering_arg.span, AtomicOrderingLoad); } else { cx.emit_spanned_lint(INVALID_ATOMIC_ORDERING, ordering_arg.span, AtomicOrderingStore); }; } } fn check_memory_fence(cx: &LateContext<'_>, expr: &Expr<'_>) { if let ExprKind::Call(ref func, ref args) = expr.kind && let ExprKind::Path(ref func_qpath) = func.kind && let Some(def_id) = cx.qpath_res(func_qpath, func.hir_id).opt_def_id() && matches!(cx.tcx.get_diagnostic_name(def_id), Some(sym::fence | sym::compiler_fence)) && Self::match_ordering(cx, &args[0]) == Some(sym::Relaxed) { cx.emit_spanned_lint(INVALID_ATOMIC_ORDERING, args[0].span, AtomicOrderingFence); } } fn check_atomic_compare_exchange(cx: &LateContext<'_>, expr: &Expr<'_>) { let Some((method, args)) = Self::inherent_atomic_method_call(cx, expr, &[sym::fetch_update, sym::compare_exchange, sym::compare_exchange_weak]) else {return }; let fail_order_arg = match method { sym::fetch_update => &args[1], sym::compare_exchange | sym::compare_exchange_weak => &args[3], _ => return, }; let Some(fail_ordering) = Self::match_ordering(cx, fail_order_arg) else { return }; if matches!(fail_ordering, sym::Release | sym::AcqRel) { cx.emit_spanned_lint( INVALID_ATOMIC_ORDERING, fail_order_arg.span, InvalidAtomicOrderingDiag { method, fail_order_arg_span: fail_order_arg.span }, ); } } } impl<'tcx> LateLintPass<'tcx> for InvalidAtomicOrdering { fn check_expr(&mut self, cx: &LateContext<'tcx>, expr: &'tcx Expr<'_>) { Self::check_atomic_load_store(cx, expr); Self::check_memory_fence(cx, expr); Self::check_atomic_compare_exchange(cx, expr); } }