summaryrefslogtreecommitdiffstats
path: root/third_party/rust/cranelift-codegen/src/machinst/abi_impl.rs
diff options
context:
space:
mode:
Diffstat (limited to 'third_party/rust/cranelift-codegen/src/machinst/abi_impl.rs')
-rw-r--r--third_party/rust/cranelift-codegen/src/machinst/abi_impl.rs1320
1 files changed, 1320 insertions, 0 deletions
diff --git a/third_party/rust/cranelift-codegen/src/machinst/abi_impl.rs b/third_party/rust/cranelift-codegen/src/machinst/abi_impl.rs
new file mode 100644
index 0000000000..e47379da37
--- /dev/null
+++ b/third_party/rust/cranelift-codegen/src/machinst/abi_impl.rs
@@ -0,0 +1,1320 @@
+//! Implementation of a vanilla ABI, shared between several machines. The
+//! implementation here assumes that arguments will be passed in registers
+//! first, then additional args on the stack; that the stack grows downward,
+//! contains a standard frame (return address and frame pointer), and the
+//! compiler is otherwise free to allocate space below that with its choice of
+//! layout; and that the machine has some notion of caller- and callee-save
+//! registers. Most modern machines, e.g. x86-64 and AArch64, should fit this
+//! mold and thus both of these backends use this shared implementation.
+//!
+//! See the documentation in specific machine backends for the "instantiation"
+//! of this generic ABI, i.e., which registers are caller/callee-save, arguments
+//! and return values, and any other special requirements.
+//!
+//! For now the implementation here assumes a 64-bit machine, but we intend to
+//! make this 32/64-bit-generic shortly.
+//!
+//! # Vanilla ABI
+//!
+//! First, arguments and return values are passed in registers up to a certain
+//! fixed count, after which they overflow onto the stack. Multiple return
+//! values either fit in registers, or are returned in a separate return-value
+//! area on the stack, given by a hidden extra parameter.
+//!
+//! Note that the exact stack layout is up to us. We settled on the
+//! below design based on several requirements. In particular, we need to be
+//! able to generate instructions (or instruction sequences) to access
+//! arguments, stack slots, and spill slots before we know how many spill slots
+//! or clobber-saves there will be, because of our pass structure. We also
+//! prefer positive offsets to negative offsets because of an asymmetry in
+//! some machines' addressing modes (e.g., on AArch64, positive offsets have a
+//! larger possible range without a long-form sequence to synthesize an
+//! arbitrary offset). Finally, it is not allowed to access memory below the
+//! current SP value.
+//!
+//! We assume that a prologue first pushes the frame pointer (and return address
+//! above that, if the machine does not do that in hardware). We set FP to point
+//! to this two-word frame record. We store all other frame slots below this
+//! two-word frame record, with the stack pointer remaining at or below this
+//! fixed frame storage for the rest of the function. We can then access frame
+//! storage slots using positive offsets from SP. In order to allow codegen for
+//! the latter before knowing how many clobber-saves we have, and also allow it
+//! while SP is being adjusted to set up a call, we implement a "nominal SP"
+//! tracking feature by which a fixup (distance between actual SP and a
+//! "nominal" SP) is known at each instruction.
+//!
+//! # Stack Layout
+//!
+//! The stack looks like:
+//!
+//! ```plain
+//! (high address)
+//!
+//! +---------------------------+
+//! | ... |
+//! | stack args |
+//! | (accessed via FP) |
+//! +---------------------------+
+//! SP at function entry -----> | return address |
+//! +---------------------------+
+//! FP after prologue --------> | FP (pushed by prologue) |
+//! +---------------------------+
+//! | ... |
+//! | spill slots |
+//! | (accessed via nominal SP) |
+//! | ... |
+//! | stack slots |
+//! | (accessed via nominal SP) |
+//! nominal SP ---------------> | (alloc'd by prologue) |
+//! +---------------------------+
+//! | ... |
+//! | clobbered callee-saves |
+//! SP at end of prologue ----> | (pushed by prologue) |
+//! +---------------------------+
+//! | [alignment as needed] |
+//! | ... |
+//! | args for call |
+//! SP before making a call --> | (pushed at callsite) |
+//! +---------------------------+
+//!
+//! (low address)
+//! ```
+//!
+//! # Multi-value Returns
+//!
+//! Note that we support multi-value returns in two ways. First, we allow for
+//! multiple return-value registers. Second, if teh appropriate flag is set, we
+//! support the SpiderMonkey Wasm ABI. For details of the multi-value return
+//! ABI, see:
+//!
+//! https://searchfox.org/mozilla-central/rev/bc3600def806859c31b2c7ac06e3d69271052a89/js/src/wasm/WasmStubs.h#134
+//!
+//! In brief:
+//! - Return values are processed in *reverse* order.
+//! - The first return value in this order (so the last return) goes into the
+//! ordinary return register.
+//! - Any further returns go in a struct-return area, allocated upwards (in
+//! address order) during the reverse traversal.
+//! - This struct-return area is provided by the caller, and a pointer to its
+//! start is passed as an invisible last (extra) argument. Normally the caller
+//! will allocate this area on the stack. When we generate calls, we place it
+//! just above the on-stack argument area.
+//! - So, for example, a function returning 4 i64's (v0, v1, v2, v3), with no
+//! formal arguments, would:
+//! - Accept a pointer `P` to the struct return area as a hidden argument in the
+//! first argument register on entry.
+//! - Return v3 in the one and only return-value register.
+//! - Return v2 in memory at `[P]`.
+//! - Return v1 in memory at `[P+8]`.
+//! - Return v0 in memory at `[P+16]`.
+
+use super::abi::*;
+use crate::binemit::StackMap;
+use crate::ir::types::*;
+use crate::ir::{ArgumentExtension, StackSlot};
+use crate::machinst::*;
+use crate::settings;
+use crate::CodegenResult;
+use crate::{ir, isa};
+use alloc::vec::Vec;
+use log::{debug, trace};
+use regalloc::{RealReg, Reg, RegClass, Set, SpillSlot, Writable};
+use std::convert::TryFrom;
+use std::marker::PhantomData;
+use std::mem;
+
+/// A location for an argument or return value.
+#[derive(Clone, Copy, Debug)]
+pub enum ABIArg {
+ /// In a real register.
+ Reg(
+ RealReg,
+ ir::Type,
+ ir::ArgumentExtension,
+ ir::ArgumentPurpose,
+ ),
+ /// Arguments only: on stack, at given offset from SP at entry.
+ Stack(i64, ir::Type, ir::ArgumentExtension, ir::ArgumentPurpose),
+}
+
+impl ABIArg {
+ /// Get the purpose of this arg.
+ fn get_purpose(self) -> ir::ArgumentPurpose {
+ match self {
+ ABIArg::Reg(_, _, _, purpose) => purpose,
+ ABIArg::Stack(_, _, _, purpose) => purpose,
+ }
+ }
+}
+
+/// Are we computing information about arguments or return values? Much of the
+/// handling is factored out into common routines; this enum allows us to
+/// distinguish which case we're handling.
+#[derive(Clone, Copy, Debug, PartialEq, Eq)]
+pub enum ArgsOrRets {
+ /// Arguments.
+ Args,
+ /// Return values.
+ Rets,
+}
+
+/// Is an instruction returned by an ABI machine-specific backend a safepoint,
+/// or not?
+#[derive(Clone, Copy, Debug, PartialEq, Eq)]
+pub enum InstIsSafepoint {
+ /// The instruction is a safepoint.
+ Yes,
+ /// The instruction is not a safepoint.
+ No,
+}
+
+/// Abstract location for a machine-specific ABI impl to translate into the
+/// appropriate addressing mode.
+#[derive(Clone, Copy, Debug)]
+pub enum StackAMode {
+ /// Offset from the frame pointer, possibly making use of a specific type
+ /// for a scaled indexing operation.
+ FPOffset(i64, ir::Type),
+ /// Offset from the nominal stack pointer, possibly making use of a specific
+ /// type for a scaled indexing operation.
+ NominalSPOffset(i64, ir::Type),
+ /// Offset from the real stack pointer, possibly making use of a specific
+ /// type for a scaled indexing operation.
+ SPOffset(i64, ir::Type),
+}
+
+/// Trait implemented by machine-specific backend to provide information about
+/// register assignments and to allow generating the specific instructions for
+/// stack loads/saves, prologues/epilogues, etc.
+pub trait ABIMachineSpec {
+ /// The instruction type.
+ type I: VCodeInst;
+
+ /// Returns the number of bits in a word, that is 32/64 for 32/64-bit architecture.
+ fn word_bits() -> u32;
+
+ /// Returns the number of bytes in a word.
+ fn word_bytes() -> u32 {
+ return Self::word_bits() / 8;
+ }
+
+ /// Returns word-size integer type.
+ fn word_type() -> Type {
+ match Self::word_bits() {
+ 32 => I32,
+ 64 => I64,
+ _ => unreachable!(),
+ }
+ }
+
+ /// Returns word register class.
+ fn word_reg_class() -> RegClass {
+ match Self::word_bits() {
+ 32 => RegClass::I32,
+ 64 => RegClass::I64,
+ _ => unreachable!(),
+ }
+ }
+
+ /// Returns required stack alignment in bytes.
+ fn stack_align(call_conv: isa::CallConv) -> u32;
+
+ /// Process a list of parameters or return values and allocate them to registers
+ /// and stack slots.
+ ///
+ /// Returns the list of argument locations, the stack-space used (rounded up
+ /// to as alignment requires), and if `add_ret_area_ptr` was passed, the
+ /// index of the extra synthetic arg that was added.
+ fn compute_arg_locs(
+ call_conv: isa::CallConv,
+ params: &[ir::AbiParam],
+ args_or_rets: ArgsOrRets,
+ add_ret_area_ptr: bool,
+ ) -> CodegenResult<(Vec<ABIArg>, i64, Option<usize>)>;
+
+ /// Returns the offset from FP to the argument area, i.e., jumping over the saved FP, return
+ /// address, and maybe other standard elements depending on ABI (e.g. Wasm TLS reg).
+ fn fp_to_arg_offset(call_conv: isa::CallConv, flags: &settings::Flags) -> i64;
+
+ /// Generate a load from the stack.
+ fn gen_load_stack(mem: StackAMode, into_reg: Writable<Reg>, ty: Type) -> Self::I;
+
+ /// Generate a store to the stack.
+ fn gen_store_stack(mem: StackAMode, from_reg: Reg, ty: Type) -> Self::I;
+
+ /// Generate a move.
+ fn gen_move(to_reg: Writable<Reg>, from_reg: Reg, ty: Type) -> Self::I;
+
+ /// Generate an integer-extend operation.
+ fn gen_extend(
+ to_reg: Writable<Reg>,
+ from_reg: Reg,
+ is_signed: bool,
+ from_bits: u8,
+ to_bits: u8,
+ ) -> Self::I;
+
+ /// Generate a return instruction.
+ fn gen_ret() -> Self::I;
+
+ /// Generate an "epilogue placeholder" instruction, recognized by lowering
+ /// when using the Baldrdash ABI.
+ fn gen_epilogue_placeholder() -> Self::I;
+
+ /// Generate an add-with-immediate. Note that even if this uses a scratch
+ /// register, it must satisfy two requirements:
+ ///
+ /// - The add-imm sequence must only clobber caller-save registers, because
+ /// it will be placed in the prologue before the clobbered callee-save
+ /// registers are saved.
+ ///
+ /// - The add-imm sequence must work correctly when `from_reg` and/or
+ /// `into_reg` are the register returned by `get_stacklimit_reg()`.
+ fn gen_add_imm(into_reg: Writable<Reg>, from_reg: Reg, imm: u32) -> SmallVec<[Self::I; 4]>;
+
+ /// Generate a sequence that traps with a `TrapCode::StackOverflow` code if
+ /// the stack pointer is less than the given limit register (assuming the
+ /// stack grows downward).
+ fn gen_stack_lower_bound_trap(limit_reg: Reg) -> SmallVec<[Self::I; 2]>;
+
+ /// Generate an instruction to compute an address of a stack slot (FP- or
+ /// SP-based offset).
+ fn gen_get_stack_addr(mem: StackAMode, into_reg: Writable<Reg>, ty: Type) -> Self::I;
+
+ /// Get a fixed register to use to compute a stack limit. This is needed for
+ /// certain sequences generated after the register allocator has already
+ /// run. This must satisfy two requirements:
+ ///
+ /// - It must be a caller-save register, because it will be clobbered in the
+ /// prologue before the clobbered callee-save registers are saved.
+ ///
+ /// - It must be safe to pass as an argument and/or destination to
+ /// `gen_add_imm()`. This is relevant when an addition with a large
+ /// immediate needs its own temporary; it cannot use the same fixed
+ /// temporary as this one.
+ fn get_stacklimit_reg() -> Reg;
+
+ /// Generate a store to the given [base+offset] address.
+ fn gen_load_base_offset(into_reg: Writable<Reg>, base: Reg, offset: i32, ty: Type) -> Self::I;
+
+ /// Generate a load from the given [base+offset] address.
+ fn gen_store_base_offset(base: Reg, offset: i32, from_reg: Reg, ty: Type) -> Self::I;
+
+ /// Adjust the stack pointer up or down.
+ fn gen_sp_reg_adjust(amount: i32) -> SmallVec<[Self::I; 2]>;
+
+ /// Generate a meta-instruction that adjusts the nominal SP offset.
+ fn gen_nominal_sp_adj(amount: i32) -> Self::I;
+
+ /// Generate the usual frame-setup sequence for this architecture: e.g.,
+ /// `push rbp / mov rbp, rsp` on x86-64, or `stp fp, lr, [sp, #-16]!` on
+ /// AArch64.
+ fn gen_prologue_frame_setup() -> SmallVec<[Self::I; 2]>;
+
+ /// Generate the usual frame-restore sequence for this architecture.
+ fn gen_epilogue_frame_restore() -> SmallVec<[Self::I; 2]>;
+
+ /// Generate a clobber-save sequence. This takes the list of *all* registers
+ /// written/modified by the function body. The implementation here is
+ /// responsible for determining which of these are callee-saved according to
+ /// the ABI. It should return a sequence of instructions that "push" or
+ /// otherwise save these values to the stack. The sequence of instructions
+ /// should adjust the stack pointer downward, and should align as necessary
+ /// according to ABI requirements.
+ ///
+ /// Returns stack bytes used as well as instructions. Does not adjust
+ /// nominal SP offset; caller will do that.
+ fn gen_clobber_save(
+ call_conv: isa::CallConv,
+ flags: &settings::Flags,
+ clobbers: &Set<Writable<RealReg>>,
+ fixed_frame_storage_size: u32,
+ outgoing_args_size: u32,
+ ) -> (u64, SmallVec<[Self::I; 16]>);
+
+ /// Generate a clobber-restore sequence. This sequence should perform the
+ /// opposite of the clobber-save sequence generated above, assuming that SP
+ /// going into the sequence is at the same point that it was left when the
+ /// clobber-save sequence finished.
+ fn gen_clobber_restore(
+ call_conv: isa::CallConv,
+ flags: &settings::Flags,
+ clobbers: &Set<Writable<RealReg>>,
+ fixed_frame_storage_size: u32,
+ outgoing_args_size: u32,
+ ) -> SmallVec<[Self::I; 16]>;
+
+ /// Generate a call instruction/sequence. This method is provided one
+ /// temporary register to use to synthesize the called address, if needed.
+ fn gen_call(
+ dest: &CallDest,
+ uses: Vec<Reg>,
+ defs: Vec<Writable<Reg>>,
+ opcode: ir::Opcode,
+ tmp: Writable<Reg>,
+ callee_conv: isa::CallConv,
+ callee_conv: isa::CallConv,
+ ) -> SmallVec<[(InstIsSafepoint, Self::I); 2]>;
+
+ /// Get the number of spillslots required for the given register-class and
+ /// type.
+ fn get_number_of_spillslots_for_value(rc: RegClass, ty: Type) -> u32;
+
+ /// Get the current virtual-SP offset from an instruction-emission state.
+ fn get_virtual_sp_offset_from_state(s: &<Self::I as MachInstEmit>::State) -> i64;
+
+ /// Get the "nominal SP to FP" offset from an instruction-emission state.
+ fn get_nominal_sp_to_fp(s: &<Self::I as MachInstEmit>::State) -> i64;
+
+ /// Get all caller-save registers, that is, registers that we expect
+ /// not to be saved across a call to a callee with the given ABI.
+ fn get_regs_clobbered_by_call(call_conv_of_callee: isa::CallConv) -> Vec<Writable<Reg>>;
+}
+
+/// ABI information shared between body (callee) and caller.
+struct ABISig {
+ /// Argument locations (regs or stack slots). Stack offsets are relative to
+ /// SP on entry to function.
+ args: Vec<ABIArg>,
+ /// Return-value locations. Stack offsets are relative to the return-area
+ /// pointer.
+ rets: Vec<ABIArg>,
+ /// Space on stack used to store arguments.
+ stack_arg_space: i64,
+ /// Space on stack used to store return values.
+ stack_ret_space: i64,
+ /// Index in `args` of the stack-return-value-area argument.
+ stack_ret_arg: Option<usize>,
+ /// Calling convention used.
+ call_conv: isa::CallConv,
+}
+
+impl ABISig {
+ fn from_func_sig<M: ABIMachineSpec>(sig: &ir::Signature) -> CodegenResult<ABISig> {
+ // Compute args and retvals from signature. Handle retvals first,
+ // because we may need to add a return-area arg to the args.
+ let (rets, stack_ret_space, _) = M::compute_arg_locs(
+ sig.call_conv,
+ &sig.returns,
+ ArgsOrRets::Rets,
+ /* extra ret-area ptr = */ false,
+ )?;
+ let need_stack_return_area = stack_ret_space > 0;
+ let (args, stack_arg_space, stack_ret_arg) = M::compute_arg_locs(
+ sig.call_conv,
+ &sig.params,
+ ArgsOrRets::Args,
+ need_stack_return_area,
+ )?;
+
+ trace!(
+ "ABISig: sig {:?} => args = {:?} rets = {:?} arg stack = {} ret stack = {} stack_ret_arg = {:?}",
+ sig,
+ args,
+ rets,
+ stack_arg_space,
+ stack_ret_space,
+ stack_ret_arg
+ );
+
+ Ok(ABISig {
+ args,
+ rets,
+ stack_arg_space,
+ stack_ret_space,
+ stack_ret_arg,
+ call_conv: sig.call_conv,
+ })
+ }
+}
+
+/// ABI object for a function body.
+pub struct ABICalleeImpl<M: ABIMachineSpec> {
+ /// Signature: arg and retval regs.
+ sig: ABISig,
+ /// Offsets to each stackslot.
+ stackslots: Vec<u32>,
+ /// Total stack size of all stackslots.
+ stackslots_size: u32,
+ /// Stack size to be reserved for outgoing arguments.
+ outgoing_args_size: u32,
+ /// Clobbered registers, from regalloc.
+ clobbered: Set<Writable<RealReg>>,
+ /// Total number of spillslots, from regalloc.
+ spillslots: Option<usize>,
+ /// Storage allocated for the fixed part of the stack frame. This is
+ /// usually the same as the total frame size below, except in the case
+ /// of the baldrdash calling convention.
+ fixed_frame_storage_size: u32,
+ /// "Total frame size", as defined by "distance between FP and nominal SP".
+ /// Some items are pushed below nominal SP, so the function may actually use
+ /// more stack than this would otherwise imply. It is simply the initial
+ /// frame/allocation size needed for stackslots and spillslots.
+ total_frame_size: Option<u32>,
+ /// The register holding the return-area pointer, if needed.
+ ret_area_ptr: Option<Writable<Reg>>,
+ /// Calling convention this function expects.
+ call_conv: isa::CallConv,
+ /// The settings controlling this function's compilation.
+ flags: settings::Flags,
+ /// Whether or not this function is a "leaf", meaning it calls no other
+ /// functions
+ is_leaf: bool,
+ /// If this function has a stack limit specified, then `Reg` is where the
+ /// stack limit will be located after the instructions specified have been
+ /// executed.
+ ///
+ /// Note that this is intended for insertion into the prologue, if
+ /// present. Also note that because the instructions here execute in the
+ /// prologue this happens after legalization/register allocation/etc so we
+ /// need to be extremely careful with each instruction. The instructions are
+ /// manually register-allocated and carefully only use caller-saved
+ /// registers and keep nothing live after this sequence of instructions.
+ stack_limit: Option<(Reg, Vec<M::I>)>,
+
+ _mach: PhantomData<M>,
+}
+
+fn get_special_purpose_param_register(
+ f: &ir::Function,
+ abi: &ABISig,
+ purpose: ir::ArgumentPurpose,
+) -> Option<Reg> {
+ let idx = f.signature.special_param_index(purpose)?;
+ match abi.args[idx] {
+ ABIArg::Reg(reg, ..) => Some(reg.to_reg()),
+ ABIArg::Stack(..) => None,
+ }
+}
+
+impl<M: ABIMachineSpec> ABICalleeImpl<M> {
+ /// Create a new body ABI instance.
+ pub fn new(f: &ir::Function, flags: settings::Flags) -> CodegenResult<Self> {
+ debug!("ABI: func signature {:?}", f.signature);
+
+ let sig = ABISig::from_func_sig::<M>(&f.signature)?;
+
+ let call_conv = f.signature.call_conv;
+ // Only these calling conventions are supported.
+ debug_assert!(
+ call_conv == isa::CallConv::SystemV
+ || call_conv == isa::CallConv::Fast
+ || call_conv == isa::CallConv::Cold
+ || call_conv.extends_baldrdash(),
+ "Unsupported calling convention: {:?}",
+ call_conv
+ );
+
+ // Compute stackslot locations and total stackslot size.
+ let mut stack_offset: u32 = 0;
+ let mut stackslots = vec![];
+ for (stackslot, data) in f.stack_slots.iter() {
+ let off = stack_offset;
+ stack_offset += data.size;
+ let mask = M::word_bytes() - 1;
+ stack_offset = (stack_offset + mask) & !mask;
+ debug_assert_eq!(stackslot.as_u32() as usize, stackslots.len());
+ stackslots.push(off);
+ }
+
+ // Figure out what instructions, if any, will be needed to check the
+ // stack limit. This can either be specified as a special-purpose
+ // argument or as a global value which often calculates the stack limit
+ // from the arguments.
+ let stack_limit =
+ get_special_purpose_param_register(f, &sig, ir::ArgumentPurpose::StackLimit)
+ .map(|reg| (reg, Vec::new()))
+ .or_else(|| f.stack_limit.map(|gv| gen_stack_limit::<M>(f, &sig, gv)));
+
+ Ok(Self {
+ sig,
+ stackslots,
+ stackslots_size: stack_offset,
+ outgoing_args_size: 0,
+ clobbered: Set::empty(),
+ spillslots: None,
+ fixed_frame_storage_size: 0,
+ total_frame_size: None,
+ ret_area_ptr: None,
+ call_conv,
+ flags,
+ is_leaf: f.is_leaf(),
+ stack_limit,
+ _mach: PhantomData,
+ })
+ }
+
+ /// Inserts instructions necessary for checking the stack limit into the
+ /// prologue.
+ ///
+ /// This function will generate instructions necessary for perform a stack
+ /// check at the header of a function. The stack check is intended to trap
+ /// if the stack pointer goes below a particular threshold, preventing stack
+ /// overflow in wasm or other code. The `stack_limit` argument here is the
+ /// register which holds the threshold below which we're supposed to trap.
+ /// This function is known to allocate `stack_size` bytes and we'll push
+ /// instructions onto `insts`.
+ ///
+ /// Note that the instructions generated here are special because this is
+ /// happening so late in the pipeline (e.g. after register allocation). This
+ /// means that we need to do manual register allocation here and also be
+ /// careful to not clobber any callee-saved or argument registers. For now
+ /// this routine makes do with the `spilltmp_reg` as one temporary
+ /// register, and a second register of `tmp2` which is caller-saved. This
+ /// should be fine for us since no spills should happen in this sequence of
+ /// instructions, so our register won't get accidentally clobbered.
+ ///
+ /// No values can be live after the prologue, but in this case that's ok
+ /// because we just need to perform a stack check before progressing with
+ /// the rest of the function.
+ fn insert_stack_check(&self, stack_limit: Reg, stack_size: u32, insts: &mut Vec<M::I>) {
+ // With no explicit stack allocated we can just emit the simple check of
+ // the stack registers against the stack limit register, and trap if
+ // it's out of bounds.
+ if stack_size == 0 {
+ insts.extend(M::gen_stack_lower_bound_trap(stack_limit));
+ return;
+ }
+
+ // Note that the 32k stack size here is pretty special. See the
+ // documentation in x86/abi.rs for why this is here. The general idea is
+ // that we're protecting against overflow in the addition that happens
+ // below.
+ if stack_size >= 32 * 1024 {
+ insts.extend(M::gen_stack_lower_bound_trap(stack_limit));
+ }
+
+ // Add the `stack_size` to `stack_limit`, placing the result in
+ // `scratch`.
+ //
+ // Note though that `stack_limit`'s register may be the same as
+ // `scratch`. If our stack size doesn't fit into an immediate this
+ // means we need a second scratch register for loading the stack size
+ // into a register.
+ let scratch = Writable::from_reg(M::get_stacklimit_reg());
+ insts.extend(M::gen_add_imm(scratch, stack_limit, stack_size).into_iter());
+ insts.extend(M::gen_stack_lower_bound_trap(scratch.to_reg()));
+ }
+}
+
+/// Generates the instructions necessary for the `gv` to be materialized into a
+/// register.
+///
+/// This function will return a register that will contain the result of
+/// evaluating `gv`. It will also return any instructions necessary to calculate
+/// the value of the register.
+///
+/// Note that global values are typically lowered to instructions via the
+/// standard legalization pass. Unfortunately though prologue generation happens
+/// so late in the pipeline that we can't use these legalization passes to
+/// generate the instructions for `gv`. As a result we duplicate some lowering
+/// of `gv` here and support only some global values. This is similar to what
+/// the x86 backend does for now, and hopefully this can be somewhat cleaned up
+/// in the future too!
+///
+/// Also note that this function will make use of `writable_spilltmp_reg()` as a
+/// temporary register to store values in if necessary. Currently after we write
+/// to this register there's guaranteed to be no spilled values between where
+/// it's used, because we're not participating in register allocation anyway!
+fn gen_stack_limit<M: ABIMachineSpec>(
+ f: &ir::Function,
+ abi: &ABISig,
+ gv: ir::GlobalValue,
+) -> (Reg, Vec<M::I>) {
+ let mut insts = Vec::new();
+ let reg = generate_gv::<M>(f, abi, gv, &mut insts);
+ return (reg, insts);
+}
+
+fn generate_gv<M: ABIMachineSpec>(
+ f: &ir::Function,
+ abi: &ABISig,
+ gv: ir::GlobalValue,
+ insts: &mut Vec<M::I>,
+) -> Reg {
+ match f.global_values[gv] {
+ // Return the direct register the vmcontext is in
+ ir::GlobalValueData::VMContext => {
+ get_special_purpose_param_register(f, abi, ir::ArgumentPurpose::VMContext)
+ .expect("no vmcontext parameter found")
+ }
+ // Load our base value into a register, then load from that register
+ // in to a temporary register.
+ ir::GlobalValueData::Load {
+ base,
+ offset,
+ global_type: _,
+ readonly: _,
+ } => {
+ let base = generate_gv::<M>(f, abi, base, insts);
+ let into_reg = Writable::from_reg(M::get_stacklimit_reg());
+ insts.push(M::gen_load_base_offset(
+ into_reg,
+ base,
+ offset.into(),
+ M::word_type(),
+ ));
+ return into_reg.to_reg();
+ }
+ ref other => panic!("global value for stack limit not supported: {}", other),
+ }
+}
+
+/// Return a type either from an optional type hint, or if not, from the default
+/// type associated with the given register's class. This is used to generate
+/// loads/spills appropriately given the type of value loaded/stored (which may
+/// be narrower than the spillslot). We usually have the type because the
+/// regalloc usually provides the vreg being spilled/reloaded, and we know every
+/// vreg's type. However, the regalloc *can* request a spill/reload without an
+/// associated vreg when needed to satisfy a safepoint (which requires all
+/// ref-typed values, even those in real registers in the original vcode, to be
+/// in spillslots).
+fn ty_from_ty_hint_or_reg_class<M: ABIMachineSpec>(r: Reg, ty: Option<Type>) -> Type {
+ match (ty, r.get_class()) {
+ // If the type is provided
+ (Some(t), _) => t,
+ // If no type is provided, this should be a register spill for a
+ // safepoint, so we only expect I32/I64 (integer) registers.
+ (None, rc) if rc == M::word_reg_class() => M::word_type(),
+ _ => panic!("Unexpected register class!"),
+ }
+}
+
+impl<M: ABIMachineSpec> ABICallee for ABICalleeImpl<M> {
+ type I = M::I;
+
+ fn temp_needed(&self) -> bool {
+ self.sig.stack_ret_arg.is_some()
+ }
+
+ fn init(&mut self, maybe_tmp: Option<Writable<Reg>>) {
+ if self.sig.stack_ret_arg.is_some() {
+ assert!(maybe_tmp.is_some());
+ self.ret_area_ptr = maybe_tmp;
+ }
+ }
+
+ fn accumulate_outgoing_args_size(&mut self, size: u32) {
+ if size > self.outgoing_args_size {
+ self.outgoing_args_size = size;
+ }
+ }
+
+ fn flags(&self) -> &settings::Flags {
+ &self.flags
+ }
+
+ fn call_conv(&self) -> isa::CallConv {
+ self.sig.call_conv
+ }
+
+ fn liveins(&self) -> Set<RealReg> {
+ let mut set: Set<RealReg> = Set::empty();
+ for &arg in &self.sig.args {
+ if let ABIArg::Reg(r, ..) = arg {
+ set.insert(r);
+ }
+ }
+ set
+ }
+
+ fn liveouts(&self) -> Set<RealReg> {
+ let mut set: Set<RealReg> = Set::empty();
+ for &ret in &self.sig.rets {
+ if let ABIArg::Reg(r, ..) = ret {
+ set.insert(r);
+ }
+ }
+ set
+ }
+
+ fn num_args(&self) -> usize {
+ self.sig.args.len()
+ }
+
+ fn num_retvals(&self) -> usize {
+ self.sig.rets.len()
+ }
+
+ fn num_stackslots(&self) -> usize {
+ self.stackslots.len()
+ }
+
+ fn gen_copy_arg_to_reg(&self, idx: usize, into_reg: Writable<Reg>) -> Self::I {
+ match &self.sig.args[idx] {
+ // Extension mode doesn't matter (we're copying out, not in; we
+ // ignore high bits by convention).
+ &ABIArg::Reg(r, ty, ..) => M::gen_move(into_reg, r.to_reg(), ty),
+ &ABIArg::Stack(off, ty, ..) => M::gen_load_stack(
+ StackAMode::FPOffset(M::fp_to_arg_offset(self.call_conv, &self.flags) + off, ty),
+ into_reg,
+ ty,
+ ),
+ }
+ }
+
+ fn arg_is_needed_in_body(&self, idx: usize) -> bool {
+ match self.sig.args[idx].get_purpose() {
+ // Special Baldrdash-specific pseudo-args that are present only to
+ // fill stack slots. Won't ever be used as ordinary values in the
+ // body.
+ ir::ArgumentPurpose::CalleeTLS | ir::ArgumentPurpose::CallerTLS => false,
+ _ => true,
+ }
+ }
+
+ fn gen_copy_reg_to_retval(&self, idx: usize, from_reg: Writable<Reg>) -> Vec<Self::I> {
+ let mut ret = Vec::new();
+ let word_bits = M::word_bits() as u8;
+ match &self.sig.rets[idx] {
+ &ABIArg::Reg(r, ty, ext, ..) => {
+ let from_bits = ty_bits(ty) as u8;
+ let dest_reg = Writable::from_reg(r.to_reg());
+ match (ext, from_bits) {
+ (ArgumentExtension::Uext, n) | (ArgumentExtension::Sext, n)
+ if n < word_bits =>
+ {
+ let signed = ext == ArgumentExtension::Sext;
+ ret.push(M::gen_extend(
+ dest_reg,
+ from_reg.to_reg(),
+ signed,
+ from_bits,
+ /* to_bits = */ word_bits,
+ ));
+ }
+ _ => ret.push(M::gen_move(dest_reg, from_reg.to_reg(), ty)),
+ };
+ }
+ &ABIArg::Stack(off, mut ty, ext, ..) => {
+ let from_bits = ty_bits(ty) as u8;
+ // A machine ABI implementation should ensure that stack frames
+ // have "reasonable" size. All current ABIs for machinst
+ // backends (aarch64 and x64) enforce a 128MB limit.
+ let off = i32::try_from(off)
+ .expect("Argument stack offset greater than 2GB; should hit impl limit first");
+ // Trash the from_reg; it should be its last use.
+ match (ext, from_bits) {
+ (ArgumentExtension::Uext, n) | (ArgumentExtension::Sext, n)
+ if n < word_bits =>
+ {
+ assert_eq!(M::word_reg_class(), from_reg.to_reg().get_class());
+ let signed = ext == ArgumentExtension::Sext;
+ ret.push(M::gen_extend(
+ from_reg,
+ from_reg.to_reg(),
+ signed,
+ from_bits,
+ /* to_bits = */ word_bits,
+ ));
+ // Store the extended version.
+ ty = M::word_type();
+ }
+ _ => {}
+ };
+ ret.push(M::gen_store_base_offset(
+ self.ret_area_ptr.unwrap().to_reg(),
+ off,
+ from_reg.to_reg(),
+ ty,
+ ));
+ }
+ }
+ ret
+ }
+
+ fn gen_retval_area_setup(&self) -> Option<Self::I> {
+ if let Some(i) = self.sig.stack_ret_arg {
+ let inst = self.gen_copy_arg_to_reg(i, self.ret_area_ptr.unwrap());
+ trace!(
+ "gen_retval_area_setup: inst {:?}; ptr reg is {:?}",
+ inst,
+ self.ret_area_ptr.unwrap().to_reg()
+ );
+ Some(inst)
+ } else {
+ trace!("gen_retval_area_setup: not needed");
+ None
+ }
+ }
+
+ fn gen_ret(&self) -> Self::I {
+ M::gen_ret()
+ }
+
+ fn gen_epilogue_placeholder(&self) -> Self::I {
+ M::gen_epilogue_placeholder()
+ }
+
+ fn set_num_spillslots(&mut self, slots: usize) {
+ self.spillslots = Some(slots);
+ }
+
+ fn set_clobbered(&mut self, clobbered: Set<Writable<RealReg>>) {
+ self.clobbered = clobbered;
+ }
+
+ /// Load from a stackslot.
+ fn load_stackslot(
+ &self,
+ slot: StackSlot,
+ offset: u32,
+ ty: Type,
+ into_reg: Writable<Reg>,
+ ) -> Self::I {
+ // Offset from beginning of stackslot area, which is at nominal SP (see
+ // [MemArg::NominalSPOffset] for more details on nominal SP tracking).
+ let stack_off = self.stackslots[slot.as_u32() as usize] as i64;
+ let sp_off: i64 = stack_off + (offset as i64);
+ trace!("load_stackslot: slot {} -> sp_off {}", slot, sp_off);
+ M::gen_load_stack(StackAMode::NominalSPOffset(sp_off, ty), into_reg, ty)
+ }
+
+ /// Store to a stackslot.
+ fn store_stackslot(&self, slot: StackSlot, offset: u32, ty: Type, from_reg: Reg) -> Self::I {
+ // Offset from beginning of stackslot area, which is at nominal SP (see
+ // [MemArg::NominalSPOffset] for more details on nominal SP tracking).
+ let stack_off = self.stackslots[slot.as_u32() as usize] as i64;
+ let sp_off: i64 = stack_off + (offset as i64);
+ trace!("store_stackslot: slot {} -> sp_off {}", slot, sp_off);
+ M::gen_store_stack(StackAMode::NominalSPOffset(sp_off, ty), from_reg, ty)
+ }
+
+ /// Produce an instruction that computes a stackslot address.
+ fn stackslot_addr(&self, slot: StackSlot, offset: u32, into_reg: Writable<Reg>) -> Self::I {
+ // Offset from beginning of stackslot area, which is at nominal SP (see
+ // [MemArg::NominalSPOffset] for more details on nominal SP tracking).
+ let stack_off = self.stackslots[slot.as_u32() as usize] as i64;
+ let sp_off: i64 = stack_off + (offset as i64);
+ M::gen_get_stack_addr(StackAMode::NominalSPOffset(sp_off, I8), into_reg, I8)
+ }
+
+ /// Load from a spillslot.
+ fn load_spillslot(&self, slot: SpillSlot, ty: Type, into_reg: Writable<Reg>) -> Self::I {
+ // Offset from beginning of spillslot area, which is at nominal SP + stackslots_size.
+ let islot = slot.get() as i64;
+ let spill_off = islot * M::word_bytes() as i64;
+ let sp_off = self.stackslots_size as i64 + spill_off;
+ trace!("load_spillslot: slot {:?} -> sp_off {}", slot, sp_off);
+ M::gen_load_stack(StackAMode::NominalSPOffset(sp_off, ty), into_reg, ty)
+ }
+
+ /// Store to a spillslot.
+ fn store_spillslot(&self, slot: SpillSlot, ty: Type, from_reg: Reg) -> Self::I {
+ // Offset from beginning of spillslot area, which is at nominal SP + stackslots_size.
+ let islot = slot.get() as i64;
+ let spill_off = islot * M::word_bytes() as i64;
+ let sp_off = self.stackslots_size as i64 + spill_off;
+ trace!("store_spillslot: slot {:?} -> sp_off {}", slot, sp_off);
+ M::gen_store_stack(StackAMode::NominalSPOffset(sp_off, ty), from_reg, ty)
+ }
+
+ fn spillslots_to_stack_map(
+ &self,
+ slots: &[SpillSlot],
+ state: &<Self::I as MachInstEmit>::State,
+ ) -> StackMap {
+ let virtual_sp_offset = M::get_virtual_sp_offset_from_state(state);
+ let nominal_sp_to_fp = M::get_nominal_sp_to_fp(state);
+ assert!(virtual_sp_offset >= 0);
+ trace!(
+ "spillslots_to_stackmap: slots = {:?}, state = {:?}",
+ slots,
+ state
+ );
+ let map_size = (virtual_sp_offset + nominal_sp_to_fp) as u32;
+ let bytes = M::word_bytes();
+ let map_words = (map_size + bytes - 1) / bytes;
+ let mut bits = std::iter::repeat(false)
+ .take(map_words as usize)
+ .collect::<Vec<bool>>();
+
+ let first_spillslot_word =
+ ((self.stackslots_size + virtual_sp_offset as u32) / bytes) as usize;
+ for &slot in slots {
+ let slot = slot.get() as usize;
+ bits[first_spillslot_word + slot] = true;
+ }
+
+ StackMap::from_slice(&bits[..])
+ }
+
+ fn gen_prologue(&mut self) -> Vec<Self::I> {
+ let mut insts = vec![];
+ if !self.call_conv.extends_baldrdash() {
+ // set up frame
+ insts.extend(M::gen_prologue_frame_setup().into_iter());
+ }
+
+ let bytes = M::word_bytes();
+ let mut total_stacksize = self.stackslots_size + bytes * self.spillslots.unwrap() as u32;
+ if self.call_conv.extends_baldrdash() {
+ debug_assert!(
+ !self.flags.enable_probestack(),
+ "baldrdash does not expect cranelift to emit stack probes"
+ );
+ total_stacksize += self.flags.baldrdash_prologue_words() as u32 * bytes;
+ }
+ let mask = M::stack_align(self.call_conv) - 1;
+ let total_stacksize = (total_stacksize + mask) & !mask; // 16-align the stack.
+
+ if !self.call_conv.extends_baldrdash() {
+ // Leaf functions with zero stack don't need a stack check if one's
+ // specified, otherwise always insert the stack check.
+ if total_stacksize > 0 || !self.is_leaf {
+ if let Some((reg, stack_limit_load)) = &self.stack_limit {
+ insts.extend_from_slice(stack_limit_load);
+ self.insert_stack_check(*reg, total_stacksize, &mut insts);
+ }
+ }
+ if total_stacksize > 0 {
+ self.fixed_frame_storage_size += total_stacksize;
+ }
+ }
+
+ // N.B.: "nominal SP", which we use to refer to stackslots and
+ // spillslots, is defined to be equal to the stack pointer at this point
+ // in the prologue.
+ //
+ // If we push any clobbers below, we emit a virtual-SP adjustment
+ // meta-instruction so that the nominal SP references behave as if SP
+ // were still at this point. See documentation for
+ // [crate::machinst::abi_impl](this module) for more details on
+ // stackframe layout and nominal SP maintenance.
+
+ // Save clobbered registers.
+ let (clobber_size, clobber_insts) = M::gen_clobber_save(
+ self.call_conv,
+ &self.flags,
+ &self.clobbered,
+ self.fixed_frame_storage_size,
+ self.outgoing_args_size,
+ );
+ insts.extend(clobber_insts);
+
+ let sp_adj = self.outgoing_args_size as i32 + clobber_size as i32;
+ if sp_adj > 0 {
+ insts.push(M::gen_nominal_sp_adj(sp_adj));
+ }
+
+ self.total_frame_size = Some(total_stacksize);
+ insts
+ }
+
+ fn gen_epilogue(&self) -> Vec<M::I> {
+ let mut insts = vec![];
+
+ // Restore clobbered registers.
+ insts.extend(M::gen_clobber_restore(
+ self.call_conv,
+ &self.flags,
+ &self.clobbered,
+ self.fixed_frame_storage_size,
+ self.outgoing_args_size,
+ ));
+
+ // N.B.: we do *not* emit a nominal SP adjustment here, because (i) there will be no
+ // references to nominal SP offsets before the return below, and (ii) the instruction
+ // emission tracks running SP offset linearly (in straight-line order), not according to
+ // the CFG, so early returns in the middle of function bodies would cause an incorrect
+ // offset for the rest of the body.
+
+ if !self.call_conv.extends_baldrdash() {
+ insts.extend(M::gen_epilogue_frame_restore());
+ insts.push(M::gen_ret());
+ }
+
+ debug!("Epilogue: {:?}", insts);
+ insts
+ }
+
+ fn frame_size(&self) -> u32 {
+ self.total_frame_size
+ .expect("frame size not computed before prologue generation")
+ }
+
+ fn stack_args_size(&self) -> u32 {
+ self.sig.stack_arg_space as u32
+ }
+
+ fn get_spillslot_size(&self, rc: RegClass, ty: Type) -> u32 {
+ M::get_number_of_spillslots_for_value(rc, ty)
+ }
+
+ fn gen_spill(&self, to_slot: SpillSlot, from_reg: RealReg, ty: Option<Type>) -> Self::I {
+ let ty = ty_from_ty_hint_or_reg_class::<M>(from_reg.to_reg(), ty);
+ self.store_spillslot(to_slot, ty, from_reg.to_reg())
+ }
+
+ fn gen_reload(
+ &self,
+ to_reg: Writable<RealReg>,
+ from_slot: SpillSlot,
+ ty: Option<Type>,
+ ) -> Self::I {
+ let ty = ty_from_ty_hint_or_reg_class::<M>(to_reg.to_reg().to_reg(), ty);
+ self.load_spillslot(from_slot, ty, to_reg.map(|r| r.to_reg()))
+ }
+
+ fn unwind_info_kind(&self) -> UnwindInfoKind {
+ match self.sig.call_conv {
+ #[cfg(feature = "unwind")]
+ isa::CallConv::Fast | isa::CallConv::Cold | isa::CallConv::SystemV => {
+ UnwindInfoKind::SystemV
+ }
+ #[cfg(feature = "unwind")]
+ isa::CallConv::WindowsFastcall => UnwindInfoKind::Windows,
+ _ => UnwindInfoKind::None,
+ }
+ }
+}
+
+fn abisig_to_uses_and_defs<M: ABIMachineSpec>(sig: &ABISig) -> (Vec<Reg>, Vec<Writable<Reg>>) {
+ // Compute uses: all arg regs.
+ let mut uses = Vec::new();
+ for arg in &sig.args {
+ match arg {
+ &ABIArg::Reg(reg, ..) => uses.push(reg.to_reg()),
+ _ => {}
+ }
+ }
+
+ // Compute defs: all retval regs, and all caller-save (clobbered) regs.
+ let mut defs = M::get_regs_clobbered_by_call(sig.call_conv);
+ for ret in &sig.rets {
+ match ret {
+ &ABIArg::Reg(reg, ..) => defs.push(Writable::from_reg(reg.to_reg())),
+ _ => {}
+ }
+ }
+
+ (uses, defs)
+}
+
+/// ABI object for a callsite.
+pub struct ABICallerImpl<M: ABIMachineSpec> {
+ /// The called function's signature.
+ sig: ABISig,
+ /// All uses for the callsite, i.e., function args.
+ uses: Vec<Reg>,
+ /// All defs for the callsite, i.e., return values and caller-saves.
+ defs: Vec<Writable<Reg>>,
+ /// Call destination.
+ dest: CallDest,
+ /// Actual call opcode; used to distinguish various types of calls.
+ opcode: ir::Opcode,
+ /// Caller's calling convention.
+ caller_conv: isa::CallConv,
+
+ _mach: PhantomData<M>,
+}
+
+/// Destination for a call.
+#[derive(Debug, Clone)]
+pub enum CallDest {
+ /// Call to an ExtName (named function symbol).
+ ExtName(ir::ExternalName, RelocDistance),
+ /// Indirect call to a function pointer in a register.
+ Reg(Reg),
+}
+
+impl<M: ABIMachineSpec> ABICallerImpl<M> {
+ /// Create a callsite ABI object for a call directly to the specified function.
+ pub fn from_func(
+ sig: &ir::Signature,
+ extname: &ir::ExternalName,
+ dist: RelocDistance,
+ caller_conv: isa::CallConv,
+ ) -> CodegenResult<ABICallerImpl<M>> {
+ let sig = ABISig::from_func_sig::<M>(sig)?;
+ let (uses, defs) = abisig_to_uses_and_defs::<M>(&sig);
+ Ok(ABICallerImpl {
+ sig,
+ uses,
+ defs,
+ dest: CallDest::ExtName(extname.clone(), dist),
+ opcode: ir::Opcode::Call,
+ caller_conv,
+ _mach: PhantomData,
+ })
+ }
+
+ /// Create a callsite ABI object for a call to a function pointer with the
+ /// given signature.
+ pub fn from_ptr(
+ sig: &ir::Signature,
+ ptr: Reg,
+ opcode: ir::Opcode,
+ caller_conv: isa::CallConv,
+ ) -> CodegenResult<ABICallerImpl<M>> {
+ let sig = ABISig::from_func_sig::<M>(sig)?;
+ let (uses, defs) = abisig_to_uses_and_defs::<M>(&sig);
+ Ok(ABICallerImpl {
+ sig,
+ uses,
+ defs,
+ dest: CallDest::Reg(ptr),
+ opcode,
+ caller_conv,
+ _mach: PhantomData,
+ })
+ }
+}
+
+fn adjust_stack_and_nominal_sp<M: ABIMachineSpec, C: LowerCtx<I = M::I>>(
+ ctx: &mut C,
+ off: i32,
+ is_sub: bool,
+) {
+ if off == 0 {
+ return;
+ }
+ let amt = if is_sub { -off } else { off };
+ for inst in M::gen_sp_reg_adjust(amt) {
+ ctx.emit(inst);
+ }
+ ctx.emit(M::gen_nominal_sp_adj(-amt));
+}
+
+impl<M: ABIMachineSpec> ABICaller for ABICallerImpl<M> {
+ type I = M::I;
+
+ fn num_args(&self) -> usize {
+ if self.sig.stack_ret_arg.is_some() {
+ self.sig.args.len() - 1
+ } else {
+ self.sig.args.len()
+ }
+ }
+
+ fn accumulate_outgoing_args_size<C: LowerCtx<I = Self::I>>(&self, ctx: &mut C) {
+ let off = self.sig.stack_arg_space + self.sig.stack_ret_space;
+ ctx.abi().accumulate_outgoing_args_size(off as u32);
+ }
+
+ fn emit_stack_pre_adjust<C: LowerCtx<I = Self::I>>(&self, ctx: &mut C) {
+ let off = self.sig.stack_arg_space + self.sig.stack_ret_space;
+ adjust_stack_and_nominal_sp::<M, C>(ctx, off as i32, /* is_sub = */ true)
+ }
+
+ fn emit_stack_post_adjust<C: LowerCtx<I = Self::I>>(&self, ctx: &mut C) {
+ let off = self.sig.stack_arg_space + self.sig.stack_ret_space;
+ adjust_stack_and_nominal_sp::<M, C>(ctx, off as i32, /* is_sub = */ false)
+ }
+
+ fn emit_copy_reg_to_arg<C: LowerCtx<I = Self::I>>(
+ &self,
+ ctx: &mut C,
+ idx: usize,
+ from_reg: Reg,
+ ) {
+ let word_rc = M::word_reg_class();
+ let word_bits = M::word_bits() as usize;
+ match &self.sig.args[idx] {
+ &ABIArg::Reg(reg, ty, ext, _)
+ if ext != ir::ArgumentExtension::None && ty_bits(ty) < word_bits =>
+ {
+ assert_eq!(word_rc, reg.get_class());
+ let signed = match ext {
+ ir::ArgumentExtension::Uext => false,
+ ir::ArgumentExtension::Sext => true,
+ _ => unreachable!(),
+ };
+ ctx.emit(M::gen_extend(
+ Writable::from_reg(reg.to_reg()),
+ from_reg,
+ signed,
+ ty_bits(ty) as u8,
+ word_bits as u8,
+ ));
+ }
+ &ABIArg::Reg(reg, ty, _, _) => {
+ ctx.emit(M::gen_move(Writable::from_reg(reg.to_reg()), from_reg, ty));
+ }
+ &ABIArg::Stack(off, mut ty, ext, _) => {
+ if ext != ir::ArgumentExtension::None && ty_bits(ty) < word_bits {
+ assert_eq!(word_rc, from_reg.get_class());
+ let signed = match ext {
+ ir::ArgumentExtension::Uext => false,
+ ir::ArgumentExtension::Sext => true,
+ _ => unreachable!(),
+ };
+ // Extend in place in the source register. Our convention is to
+ // treat high bits as undefined for values in registers, so this
+ // is safe, even for an argument that is nominally read-only.
+ ctx.emit(M::gen_extend(
+ Writable::from_reg(from_reg),
+ from_reg,
+ signed,
+ ty_bits(ty) as u8,
+ word_bits as u8,
+ ));
+ // Store the extended version.
+ ty = M::word_type();
+ }
+ ctx.emit(M::gen_store_stack(
+ StackAMode::SPOffset(off, ty),
+ from_reg,
+ ty,
+ ));
+ }
+ }
+ }
+
+ fn emit_copy_retval_to_reg<C: LowerCtx<I = Self::I>>(
+ &self,
+ ctx: &mut C,
+ idx: usize,
+ into_reg: Writable<Reg>,
+ ) {
+ match &self.sig.rets[idx] {
+ // Extension mode doesn't matter because we're copying out, not in,
+ // and we ignore high bits in our own registers by convention.
+ &ABIArg::Reg(reg, ty, _, _) => ctx.emit(M::gen_move(into_reg, reg.to_reg(), ty)),
+ &ABIArg::Stack(off, ty, _, _) => {
+ let ret_area_base = self.sig.stack_arg_space;
+ ctx.emit(M::gen_load_stack(
+ StackAMode::SPOffset(off + ret_area_base, ty),
+ into_reg,
+ ty,
+ ));
+ }
+ }
+ }
+
+ fn emit_call<C: LowerCtx<I = Self::I>>(&mut self, ctx: &mut C) {
+ let (uses, defs) = (
+ mem::replace(&mut self.uses, Default::default()),
+ mem::replace(&mut self.defs, Default::default()),
+ );
+ let word_rc = M::word_reg_class();
+ let word_type = M::word_type();
+ if let Some(i) = self.sig.stack_ret_arg {
+ let rd = ctx.alloc_tmp(word_rc, word_type);
+ let ret_area_base = self.sig.stack_arg_space;
+ ctx.emit(M::gen_get_stack_addr(
+ StackAMode::SPOffset(ret_area_base, I8),
+ rd,
+ I8,
+ ));
+ self.emit_copy_reg_to_arg(ctx, i, rd.to_reg());
+ }
+ let tmp = ctx.alloc_tmp(word_rc, word_type);
+ for (is_safepoint, inst) in M::gen_call(
+ &self.dest,
+ uses,
+ defs,
+ self.opcode,
+ tmp,
+ self.sig.call_conv,
+ self.caller_conv,
+ )
+ .into_iter()
+ {
+ match is_safepoint {
+ InstIsSafepoint::Yes => ctx.emit_safepoint(inst),
+ InstIsSafepoint::No => ctx.emit(inst),
+ }
+ }
+ }
+}