cranelift_codegen/machinst/
abi.rs

1//! Implementation of a vanilla ABI, shared between several machines. The
2//! implementation here assumes that arguments will be passed in registers
3//! first, then additional args on the stack; that the stack grows downward,
4//! contains a standard frame (return address and frame pointer), and the
5//! compiler is otherwise free to allocate space below that with its choice of
6//! layout; and that the machine has some notion of caller- and callee-save
7//! registers. Most modern machines, e.g. x86-64 and AArch64, should fit this
8//! mold and thus both of these backends use this shared implementation.
9//!
10//! See the documentation in specific machine backends for the "instantiation"
11//! of this generic ABI, i.e., which registers are caller/callee-save, arguments
12//! and return values, and any other special requirements.
13//!
14//! For now the implementation here assumes a 64-bit machine, but we intend to
15//! make this 32/64-bit-generic shortly.
16//!
17//! # Vanilla ABI
18//!
19//! First, arguments and return values are passed in registers up to a certain
20//! fixed count, after which they overflow onto the stack. Multiple return
21//! values either fit in registers, or are returned in a separate return-value
22//! area on the stack, given by a hidden extra parameter.
23//!
24//! Note that the exact stack layout is up to us. We settled on the
25//! below design based on several requirements. In particular, we need
26//! to be able to generate instructions (or instruction sequences) to
27//! access arguments, stack slots, and spill slots before we know how
28//! many spill slots or clobber-saves there will be, because of our
29//! pass structure. We also prefer positive offsets to negative
30//! offsets because of an asymmetry in some machines' addressing modes
31//! (e.g., on AArch64, positive offsets have a larger possible range
32//! without a long-form sequence to synthesize an arbitrary
33//! offset). We also need clobber-save registers to be "near" the
34//! frame pointer: Windows unwind information requires it to be within
35//! 240 bytes of RBP. Finally, it is not allowed to access memory
36//! below the current SP value.
37//!
38//! We assume that a prologue first pushes the frame pointer (and
39//! return address above that, if the machine does not do that in
40//! hardware). We set FP to point to this two-word frame record. We
41//! store all other frame slots below this two-word frame record, as
42//! well as enough space for arguments to the largest possible
43//! function call. The stack pointer then remains at this position
44//! for the duration of the function, allowing us to address all
45//! frame storage at positive offsets from SP.
46//!
47//! Note that if we ever support dynamic stack-space allocation (for
48//! `alloca`), we will need a way to reference spill slots and stack
49//! slots relative to a dynamic SP, because we will no longer be able
50//! to know a static offset from SP to the slots at any particular
51//! program point. Probably the best solution at that point will be to
52//! revert to using the frame pointer as the reference for all slots,
53//! to allow generating spill/reload and stackslot accesses before we
54//! know how large the clobber-saves will be.
55//!
56//! # Stack Layout
57//!
58//! The stack looks like:
59//!
60//! ```plain
61//!   (high address)
62//!                              |          ...              |
63//!                              | caller frames             |
64//!                              |          ...              |
65//!                              +===========================+
66//!                              |          ...              |
67//!                              | stack args                |
68//! Canonical Frame Address -->  | (accessed via FP)         |
69//!                              +---------------------------+
70//! SP at function entry ----->  | return address            |
71//!                              +---------------------------+
72//! FP after prologue -------->  | FP (pushed by prologue)   |
73//!                              +---------------------------+           -----
74//!                              |          ...              |             |
75//!                              | clobbered callee-saves    |             |
76//! unwind-frame base -------->  | (pushed by prologue)      |             |
77//!                              +---------------------------+   -----     |
78//!                              |          ...              |     |       |
79//!                              | spill slots               |     |       |
80//!                              | (accessed via SP)         |   fixed   active
81//!                              |          ...              |   frame    size
82//!                              | stack slots               |  storage    |
83//!                              | (accessed via SP)         |    size     |
84//!                              | (alloc'd by prologue)     |     |       |
85//!                              +---------------------------+   -----     |
86//!                              | [alignment as needed]     |             |
87//!                              |          ...              |             |
88//!                              | args for largest call     |             |
89//! SP ----------------------->  | (alloc'd by prologue)     |             |
90//!                              +===========================+           -----
91//!
92//!   (low address)
93//! ```
94//!
95//! # Multi-value Returns
96//!
97//! We support multi-value returns by using multiple return-value
98//! registers. In some cases this is an extension of the base system
99//! ABI. See each platform's `abi.rs` implementation for details.
100
101use crate::CodegenError;
102use crate::entity::SecondaryMap;
103use crate::ir::types::*;
104use crate::ir::{ArgumentExtension, ArgumentPurpose, ExceptionTag, Signature};
105use crate::isa::TargetIsa;
106use crate::settings::ProbestackStrategy;
107use crate::{ir, isa};
108use crate::{machinst::*, trace};
109use alloc::boxed::Box;
110use cranelift_entity::packed_option::PackedOption;
111use regalloc2::{MachineEnv, PReg, PRegSet};
112use rustc_hash::FxHashMap;
113use smallvec::smallvec;
114use std::collections::HashMap;
115use std::marker::PhantomData;
116
117/// A small vector of instructions (with some reasonable size); appropriate for
118/// a small fixed sequence implementing one operation.
119pub type SmallInstVec<I> = SmallVec<[I; 4]>;
120
121/// A type used by backends to track argument-binding info in the "args"
122/// pseudoinst. The pseudoinst holds a vec of `ArgPair` structs.
123#[derive(Clone, Debug)]
124pub struct ArgPair {
125    /// The vreg that is defined by this args pseudoinst.
126    pub vreg: Writable<Reg>,
127    /// The preg that the arg arrives in; this constrains the vreg's
128    /// placement at the pseudoinst.
129    pub preg: Reg,
130}
131
132/// A type used by backends to track return register binding info in the "ret"
133/// pseudoinst. The pseudoinst holds a vec of `RetPair` structs.
134#[derive(Clone, Debug)]
135pub struct RetPair {
136    /// The vreg that is returned by this pseudionst.
137    pub vreg: Reg,
138    /// The preg that the arg is returned through; this constrains the vreg's
139    /// placement at the pseudoinst.
140    pub preg: Reg,
141}
142
143/// A location for (part of) an argument or return value. These "storage slots"
144/// are specified for each register-sized part of an argument.
145#[derive(Clone, Copy, Debug, PartialEq, Eq)]
146pub enum ABIArgSlot {
147    /// In a real register.
148    Reg {
149        /// Register that holds this arg.
150        reg: RealReg,
151        /// Value type of this arg.
152        ty: ir::Type,
153        /// Should this arg be zero- or sign-extended?
154        extension: ir::ArgumentExtension,
155    },
156    /// Arguments only: on stack, at given offset from SP at entry.
157    Stack {
158        /// Offset of this arg relative to the base of stack args.
159        offset: i64,
160        /// Value type of this arg.
161        ty: ir::Type,
162        /// Should this arg be zero- or sign-extended?
163        extension: ir::ArgumentExtension,
164    },
165}
166
167impl ABIArgSlot {
168    /// The type of the value that will be stored in this slot.
169    pub fn get_type(&self) -> ir::Type {
170        match self {
171            ABIArgSlot::Reg { ty, .. } => *ty,
172            ABIArgSlot::Stack { ty, .. } => *ty,
173        }
174    }
175}
176
177/// A vector of `ABIArgSlot`s. Inline capacity for one element because basically
178/// 100% of values use one slot. Only `i128`s need multiple slots, and they are
179/// super rare (and never happen with Wasm).
180pub type ABIArgSlotVec = SmallVec<[ABIArgSlot; 1]>;
181
182/// An ABIArg is composed of one or more parts. This allows for a CLIF-level
183/// Value to be passed with its parts in more than one location at the ABI
184/// level. For example, a 128-bit integer may be passed in two 64-bit registers,
185/// or even a 64-bit register and a 64-bit stack slot, on a 64-bit machine. The
186/// number of "parts" should correspond to the number of registers used to store
187/// this type according to the machine backend.
188///
189/// As an invariant, the `purpose` for every part must match. As a further
190/// invariant, a `StructArg` part cannot appear with any other part.
191#[derive(Clone, Debug)]
192pub enum ABIArg {
193    /// Storage slots (registers or stack locations) for each part of the
194    /// argument value. The number of slots must equal the number of register
195    /// parts used to store a value of this type.
196    Slots {
197        /// Slots, one per register part.
198        slots: ABIArgSlotVec,
199        /// Purpose of this arg.
200        purpose: ir::ArgumentPurpose,
201    },
202    /// Structure argument. We reserve stack space for it, but the CLIF-level
203    /// semantics are a little weird: the value passed to the call instruction,
204    /// and received in the corresponding block param, is a *pointer*. On the
205    /// caller side, we memcpy the data from the passed-in pointer to the stack
206    /// area; on the callee side, we compute a pointer to this stack area and
207    /// provide that as the argument's value.
208    StructArg {
209        /// Offset of this arg relative to base of stack args.
210        offset: i64,
211        /// Size of this arg on the stack.
212        size: u64,
213        /// Purpose of this arg.
214        purpose: ir::ArgumentPurpose,
215    },
216    /// Implicit argument. Similar to a StructArg, except that we have the
217    /// target type, not a pointer type, at the CLIF-level. This argument is
218    /// still being passed via reference implicitly.
219    ImplicitPtrArg {
220        /// Register or stack slot holding a pointer to the buffer.
221        pointer: ABIArgSlot,
222        /// Offset of the argument buffer.
223        offset: i64,
224        /// Type of the implicit argument.
225        ty: Type,
226        /// Purpose of this arg.
227        purpose: ir::ArgumentPurpose,
228    },
229}
230
231impl ABIArg {
232    /// Create an ABIArg from one register.
233    pub fn reg(
234        reg: RealReg,
235        ty: ir::Type,
236        extension: ir::ArgumentExtension,
237        purpose: ir::ArgumentPurpose,
238    ) -> ABIArg {
239        ABIArg::Slots {
240            slots: smallvec![ABIArgSlot::Reg { reg, ty, extension }],
241            purpose,
242        }
243    }
244
245    /// Create an ABIArg from one stack slot.
246    pub fn stack(
247        offset: i64,
248        ty: ir::Type,
249        extension: ir::ArgumentExtension,
250        purpose: ir::ArgumentPurpose,
251    ) -> ABIArg {
252        ABIArg::Slots {
253            slots: smallvec![ABIArgSlot::Stack {
254                offset,
255                ty,
256                extension,
257            }],
258            purpose,
259        }
260    }
261}
262
263/// Are we computing information about arguments or return values? Much of the
264/// handling is factored out into common routines; this enum allows us to
265/// distinguish which case we're handling.
266#[derive(Clone, Copy, Debug, PartialEq, Eq)]
267pub enum ArgsOrRets {
268    /// Arguments.
269    Args,
270    /// Return values.
271    Rets,
272}
273
274/// Abstract location for a machine-specific ABI impl to translate into the
275/// appropriate addressing mode.
276#[derive(Clone, Copy, Debug, PartialEq, Eq)]
277pub enum StackAMode {
278    /// Offset into the current frame's argument area.
279    IncomingArg(i64, u32),
280    /// Offset within the stack slots in the current frame.
281    Slot(i64),
282    /// Offset into the callee frame's argument area.
283    OutgoingArg(i64),
284}
285
286impl StackAMode {
287    fn offset_by(&self, offset: u32) -> Self {
288        match self {
289            StackAMode::IncomingArg(off, size) => {
290                StackAMode::IncomingArg(off.checked_add(i64::from(offset)).unwrap(), *size)
291            }
292            StackAMode::Slot(off) => StackAMode::Slot(off.checked_add(i64::from(offset)).unwrap()),
293            StackAMode::OutgoingArg(off) => {
294                StackAMode::OutgoingArg(off.checked_add(i64::from(offset)).unwrap())
295            }
296        }
297    }
298}
299
300/// Trait implemented by machine-specific backend to represent ISA flags.
301pub trait IsaFlags: Clone {
302    /// Get a flag indicating whether forward-edge CFI is enabled.
303    fn is_forward_edge_cfi_enabled(&self) -> bool {
304        false
305    }
306}
307
308/// Used as an out-parameter to accumulate a sequence of `ABIArg`s in
309/// `ABIMachineSpec::compute_arg_locs`. Wraps the shared allocation for all
310/// `ABIArg`s in `SigSet` and exposes just the args for the current
311/// `compute_arg_locs` call.
312pub struct ArgsAccumulator<'a> {
313    sig_set_abi_args: &'a mut Vec<ABIArg>,
314    start: usize,
315    non_formal_flag: bool,
316}
317
318impl<'a> ArgsAccumulator<'a> {
319    fn new(sig_set_abi_args: &'a mut Vec<ABIArg>) -> Self {
320        let start = sig_set_abi_args.len();
321        ArgsAccumulator {
322            sig_set_abi_args,
323            start,
324            non_formal_flag: false,
325        }
326    }
327
328    #[inline]
329    pub fn push(&mut self, arg: ABIArg) {
330        debug_assert!(!self.non_formal_flag);
331        self.sig_set_abi_args.push(arg)
332    }
333
334    #[inline]
335    pub fn push_non_formal(&mut self, arg: ABIArg) {
336        self.non_formal_flag = true;
337        self.sig_set_abi_args.push(arg)
338    }
339
340    #[inline]
341    pub fn args(&self) -> &[ABIArg] {
342        &self.sig_set_abi_args[self.start..]
343    }
344
345    #[inline]
346    pub fn args_mut(&mut self) -> &mut [ABIArg] {
347        &mut self.sig_set_abi_args[self.start..]
348    }
349}
350
351/// Trait implemented by machine-specific backend to provide information about
352/// register assignments and to allow generating the specific instructions for
353/// stack loads/saves, prologues/epilogues, etc.
354pub trait ABIMachineSpec {
355    /// The instruction type.
356    type I: VCodeInst;
357
358    /// The ISA flags type.
359    type F: IsaFlags;
360
361    /// This is the limit for the size of argument and return-value areas on the
362    /// stack. We place a reasonable limit here to avoid integer overflow issues
363    /// with 32-bit arithmetic.
364    const STACK_ARG_RET_SIZE_LIMIT: u32;
365
366    /// Returns the number of bits in a word, that is 32/64 for 32/64-bit architecture.
367    fn word_bits() -> u32;
368
369    /// Returns the number of bytes in a word.
370    fn word_bytes() -> u32 {
371        return Self::word_bits() / 8;
372    }
373
374    /// Returns word-size integer type.
375    fn word_type() -> Type {
376        match Self::word_bits() {
377            32 => I32,
378            64 => I64,
379            _ => unreachable!(),
380        }
381    }
382
383    /// Returns word register class.
384    fn word_reg_class() -> RegClass {
385        RegClass::Int
386    }
387
388    /// Returns required stack alignment in bytes.
389    fn stack_align(call_conv: isa::CallConv) -> u32;
390
391    /// Process a list of parameters or return values and allocate them to registers
392    /// and stack slots.
393    ///
394    /// The argument locations should be pushed onto the given `ArgsAccumulator`
395    /// in order. Any extra arguments added (such as return area pointers)
396    /// should come at the end of the list so that the first N lowered
397    /// parameters align with the N clif parameters.
398    ///
399    /// Returns the stack-space used (rounded up to as alignment requires), and
400    /// if `add_ret_area_ptr` was passed, the index of the extra synthetic arg
401    /// that was added.
402    fn compute_arg_locs(
403        call_conv: isa::CallConv,
404        flags: &settings::Flags,
405        params: &[ir::AbiParam],
406        args_or_rets: ArgsOrRets,
407        add_ret_area_ptr: bool,
408        args: ArgsAccumulator,
409    ) -> CodegenResult<(u32, Option<usize>)>;
410
411    /// Generate a load from the stack.
412    fn gen_load_stack(mem: StackAMode, into_reg: Writable<Reg>, ty: Type) -> Self::I;
413
414    /// Generate a store to the stack.
415    fn gen_store_stack(mem: StackAMode, from_reg: Reg, ty: Type) -> Self::I;
416
417    /// Generate a move.
418    fn gen_move(to_reg: Writable<Reg>, from_reg: Reg, ty: Type) -> Self::I;
419
420    /// Generate an integer-extend operation.
421    fn gen_extend(
422        to_reg: Writable<Reg>,
423        from_reg: Reg,
424        is_signed: bool,
425        from_bits: u8,
426        to_bits: u8,
427    ) -> Self::I;
428
429    /// Generate an "args" pseudo-instruction to capture input args in
430    /// registers.
431    fn gen_args(args: Vec<ArgPair>) -> Self::I;
432
433    /// Generate a "rets" pseudo-instruction that moves vregs to return
434    /// registers.
435    fn gen_rets(rets: Vec<RetPair>) -> Self::I;
436
437    /// Generate an add-with-immediate. Note that even if this uses a scratch
438    /// register, it must satisfy two requirements:
439    ///
440    /// - The add-imm sequence must only clobber caller-save registers that are
441    ///   not used for arguments, because it will be placed in the prologue
442    ///   before the clobbered callee-save registers are saved.
443    ///
444    /// - The add-imm sequence must work correctly when `from_reg` and/or
445    ///   `into_reg` are the register returned by `get_stacklimit_reg()`.
446    fn gen_add_imm(
447        call_conv: isa::CallConv,
448        into_reg: Writable<Reg>,
449        from_reg: Reg,
450        imm: u32,
451    ) -> SmallInstVec<Self::I>;
452
453    /// Generate a sequence that traps with a `TrapCode::StackOverflow` code if
454    /// the stack pointer is less than the given limit register (assuming the
455    /// stack grows downward).
456    fn gen_stack_lower_bound_trap(limit_reg: Reg) -> SmallInstVec<Self::I>;
457
458    /// Generate an instruction to compute an address of a stack slot (FP- or
459    /// SP-based offset).
460    fn gen_get_stack_addr(mem: StackAMode, into_reg: Writable<Reg>) -> Self::I;
461
462    /// Get a fixed register to use to compute a stack limit. This is needed for
463    /// certain sequences generated after the register allocator has already
464    /// run. This must satisfy two requirements:
465    ///
466    /// - It must be a caller-save register that is not used for arguments,
467    ///   because it will be clobbered in the prologue before the clobbered
468    ///   callee-save registers are saved.
469    ///
470    /// - It must be safe to pass as an argument and/or destination to
471    ///   `gen_add_imm()`. This is relevant when an addition with a large
472    ///   immediate needs its own temporary; it cannot use the same fixed
473    ///   temporary as this one.
474    fn get_stacklimit_reg(call_conv: isa::CallConv) -> Reg;
475
476    /// Generate a load to the given [base+offset] address.
477    fn gen_load_base_offset(into_reg: Writable<Reg>, base: Reg, offset: i32, ty: Type) -> Self::I;
478
479    /// Generate a store from the given [base+offset] address.
480    fn gen_store_base_offset(base: Reg, offset: i32, from_reg: Reg, ty: Type) -> Self::I;
481
482    /// Adjust the stack pointer up or down.
483    fn gen_sp_reg_adjust(amount: i32) -> SmallInstVec<Self::I>;
484
485    /// Compute a FrameLayout structure containing a sorted list of all clobbered
486    /// registers that are callee-saved according to the ABI, as well as the sizes
487    /// of all parts of the stack frame.  The result is used to emit the prologue
488    /// and epilogue routines.
489    fn compute_frame_layout(
490        call_conv: isa::CallConv,
491        flags: &settings::Flags,
492        sig: &Signature,
493        regs: &[Writable<RealReg>],
494        is_leaf: bool,
495        incoming_args_size: u32,
496        tail_args_size: u32,
497        stackslots_size: u32,
498        fixed_frame_storage_size: u32,
499        outgoing_args_size: u32,
500    ) -> FrameLayout;
501
502    /// Generate the usual frame-setup sequence for this architecture: e.g.,
503    /// `push rbp / mov rbp, rsp` on x86-64, or `stp fp, lr, [sp, #-16]!` on
504    /// AArch64.
505    fn gen_prologue_frame_setup(
506        call_conv: isa::CallConv,
507        flags: &settings::Flags,
508        isa_flags: &Self::F,
509        frame_layout: &FrameLayout,
510    ) -> SmallInstVec<Self::I>;
511
512    /// Generate the usual frame-restore sequence for this architecture.
513    fn gen_epilogue_frame_restore(
514        call_conv: isa::CallConv,
515        flags: &settings::Flags,
516        isa_flags: &Self::F,
517        frame_layout: &FrameLayout,
518    ) -> SmallInstVec<Self::I>;
519
520    /// Generate a return instruction.
521    fn gen_return(
522        call_conv: isa::CallConv,
523        isa_flags: &Self::F,
524        frame_layout: &FrameLayout,
525    ) -> SmallInstVec<Self::I>;
526
527    /// Generate a probestack call.
528    fn gen_probestack(insts: &mut SmallInstVec<Self::I>, frame_size: u32);
529
530    /// Generate a inline stack probe.
531    fn gen_inline_probestack(
532        insts: &mut SmallInstVec<Self::I>,
533        call_conv: isa::CallConv,
534        frame_size: u32,
535        guard_size: u32,
536    );
537
538    /// Generate a clobber-save sequence. The implementation here should return
539    /// a sequence of instructions that "push" or otherwise save to the stack all
540    /// registers written/modified by the function body that are callee-saved.
541    /// The sequence of instructions should adjust the stack pointer downward,
542    /// and should align as necessary according to ABI requirements.
543    fn gen_clobber_save(
544        call_conv: isa::CallConv,
545        flags: &settings::Flags,
546        frame_layout: &FrameLayout,
547    ) -> SmallVec<[Self::I; 16]>;
548
549    /// Generate a clobber-restore sequence. This sequence should perform the
550    /// opposite of the clobber-save sequence generated above, assuming that SP
551    /// going into the sequence is at the same point that it was left when the
552    /// clobber-save sequence finished.
553    fn gen_clobber_restore(
554        call_conv: isa::CallConv,
555        flags: &settings::Flags,
556        frame_layout: &FrameLayout,
557    ) -> SmallVec<[Self::I; 16]>;
558
559    /// Generate a memcpy invocation. Used to set up struct
560    /// args. Takes `src`, `dst` as read-only inputs and passes a temporary
561    /// allocator.
562    fn gen_memcpy<F: FnMut(Type) -> Writable<Reg>>(
563        call_conv: isa::CallConv,
564        dst: Reg,
565        src: Reg,
566        size: usize,
567        alloc_tmp: F,
568    ) -> SmallVec<[Self::I; 8]>;
569
570    /// Get the number of spillslots required for the given register-class.
571    fn get_number_of_spillslots_for_value(
572        rc: RegClass,
573        target_vector_bytes: u32,
574        isa_flags: &Self::F,
575    ) -> u32;
576
577    /// Get the ABI-dependent MachineEnv for managing register allocation.
578    fn get_machine_env(flags: &settings::Flags, call_conv: isa::CallConv) -> &MachineEnv;
579
580    /// Get all caller-save registers, that is, registers that we expect
581    /// not to be saved across a call to a callee with the given ABI.
582    fn get_regs_clobbered_by_call(
583        call_conv_of_callee: isa::CallConv,
584        is_exception: bool,
585    ) -> PRegSet;
586
587    /// Get the needed extension mode, given the mode attached to the argument
588    /// in the signature and the calling convention. The input (the attribute in
589    /// the signature) specifies what extension type should be done *if* the ABI
590    /// requires extension to the full register; this method's return value
591    /// indicates whether the extension actually *will* be done.
592    fn get_ext_mode(
593        call_conv: isa::CallConv,
594        specified: ir::ArgumentExtension,
595    ) -> ir::ArgumentExtension;
596
597    /// Get a temporary register that is available to use after a call
598    /// completes and that does not interfere with register-carried
599    /// return values. This is used to move stack-carried return
600    /// values directly into spillslots if needed.
601    fn retval_temp_reg(call_conv_of_callee: isa::CallConv) -> Writable<Reg>;
602
603    /// Get the exception payload registers, if any, for a calling
604    /// convention.
605    fn exception_payload_regs(_call_conv: isa::CallConv) -> &'static [Reg] {
606        &[]
607    }
608}
609
610/// Out-of-line data for calls, to keep the size of `Inst` down.
611#[derive(Clone, Debug)]
612pub struct CallInfo<T> {
613    /// Receiver of this call
614    pub dest: T,
615    /// Register uses of this call.
616    pub uses: CallArgList,
617    /// Register defs of this call.
618    pub defs: CallRetList,
619    /// Registers clobbered by this call, as per its calling convention.
620    pub clobbers: PRegSet,
621    /// The calling convention of the callee.
622    pub callee_conv: isa::CallConv,
623    /// The calling convention of the caller.
624    pub caller_conv: isa::CallConv,
625    /// The number of bytes that the callee will pop from the stack for the
626    /// caller, if any. (Used for popping stack arguments with the `tail`
627    /// calling convention.)
628    pub callee_pop_size: u32,
629    /// Information for a try-call, if this is one. We combine
630    /// handling of calls and try-calls as much as possible to share
631    /// argument/return logic; they mostly differ in the metadata that
632    /// they emit, which this information feeds into.
633    pub try_call_info: Option<TryCallInfo>,
634}
635
636/// Out-of-line information present on `try_call` instructions only:
637/// information that is used to generate exception-handling tables and
638/// link up to destination blocks properly.
639#[derive(Clone, Debug)]
640pub struct TryCallInfo {
641    /// The target to jump to on a normal returhn.
642    pub continuation: MachLabel,
643    /// Exception tags to catch and corresponding destination labels.
644    pub exception_dests: Box<[(PackedOption<ExceptionTag>, MachLabel)]>,
645}
646
647impl<T> CallInfo<T> {
648    /// Creates an empty set of info with no clobbers/uses/etc with the
649    /// specified ABI
650    pub fn empty(dest: T, call_conv: isa::CallConv) -> CallInfo<T> {
651        CallInfo {
652            dest,
653            uses: smallvec![],
654            defs: smallvec![],
655            clobbers: PRegSet::empty(),
656            caller_conv: call_conv,
657            callee_conv: call_conv,
658            callee_pop_size: 0,
659            try_call_info: None,
660        }
661    }
662}
663
664/// The id of an ABI signature within the `SigSet`.
665#[derive(Copy, Clone, PartialEq, Eq, Hash, PartialOrd, Ord)]
666pub struct Sig(u32);
667cranelift_entity::entity_impl!(Sig);
668
669impl Sig {
670    fn prev(self) -> Option<Sig> {
671        self.0.checked_sub(1).map(Sig)
672    }
673}
674
675/// ABI information shared between body (callee) and caller.
676#[derive(Clone, Debug)]
677pub struct SigData {
678    /// Currently both return values and arguments are stored in a continuous space vector
679    /// in `SigSet::abi_args`.
680    ///
681    /// ```plain
682    ///                  +----------------------------------------------+
683    ///                  | return values                                |
684    ///                  | ...                                          |
685    ///   rets_end   --> +----------------------------------------------+
686    ///                  | arguments                                    |
687    ///                  | ...                                          |
688    ///   args_end   --> +----------------------------------------------+
689    ///
690    /// ```
691    ///
692    /// Note we only store two offsets as rets_end == args_start, and rets_start == prev.args_end.
693    ///
694    /// Argument location ending offset (regs or stack slots). Stack offsets are relative to
695    /// SP on entry to function.
696    ///
697    /// This is a index into the `SigSet::abi_args`.
698    args_end: u32,
699
700    /// Return-value location ending offset. Stack offsets are relative to the return-area
701    /// pointer.
702    ///
703    /// This is a index into the `SigSet::abi_args`.
704    rets_end: u32,
705
706    /// Space on stack used to store arguments. We're storing the size in u32 to
707    /// reduce the size of the struct.
708    sized_stack_arg_space: u32,
709
710    /// Space on stack used to store return values. We're storing the size in u32 to
711    /// reduce the size of the struct.
712    sized_stack_ret_space: u32,
713
714    /// Index in `args` of the stack-return-value-area argument.
715    stack_ret_arg: Option<u16>,
716
717    /// Calling convention used.
718    call_conv: isa::CallConv,
719}
720
721impl SigData {
722    /// Get total stack space required for arguments.
723    pub fn sized_stack_arg_space(&self) -> u32 {
724        self.sized_stack_arg_space
725    }
726
727    /// Get total stack space required for return values.
728    pub fn sized_stack_ret_space(&self) -> u32 {
729        self.sized_stack_ret_space
730    }
731
732    /// Get calling convention used.
733    pub fn call_conv(&self) -> isa::CallConv {
734        self.call_conv
735    }
736
737    /// The index of the stack-return-value-area argument, if any.
738    pub fn stack_ret_arg(&self) -> Option<u16> {
739        self.stack_ret_arg
740    }
741}
742
743/// A (mostly) deduplicated set of ABI signatures.
744///
745/// We say "mostly" because we do not dedupe between signatures interned via
746/// `ir::SigRef` (direct and indirect calls; the vast majority of signatures in
747/// this set) vs via `ir::Signature` (the callee itself and libcalls). Doing
748/// this final bit of deduplication would require filling out the
749/// `ir_signature_to_abi_sig`, which is a bunch of allocations (not just the
750/// hash map itself but params and returns vecs in each signature) that we want
751/// to avoid.
752///
753/// In general, prefer using the `ir::SigRef`-taking methods to the
754/// `ir::Signature`-taking methods when you can get away with it, as they don't
755/// require cloning non-copy types that will trigger heap allocations.
756///
757/// This type can be indexed by `Sig` to access its associated `SigData`.
758pub struct SigSet {
759    /// Interned `ir::Signature`s that we already have an ABI signature for.
760    ir_signature_to_abi_sig: FxHashMap<ir::Signature, Sig>,
761
762    /// Interned `ir::SigRef`s that we already have an ABI signature for.
763    ir_sig_ref_to_abi_sig: SecondaryMap<ir::SigRef, Option<Sig>>,
764
765    /// A single, shared allocation for all `ABIArg`s used by all
766    /// `SigData`s. Each `SigData` references its args/rets via indices into
767    /// this allocation.
768    abi_args: Vec<ABIArg>,
769
770    /// The actual ABI signatures, keyed by `Sig`.
771    sigs: PrimaryMap<Sig, SigData>,
772}
773
774impl SigSet {
775    /// Construct a new `SigSet`, interning all of the signatures used by the
776    /// given function.
777    pub fn new<M>(func: &ir::Function, flags: &settings::Flags) -> CodegenResult<Self>
778    where
779        M: ABIMachineSpec,
780    {
781        let arg_estimate = func.dfg.signatures.len() * 6;
782
783        let mut sigs = SigSet {
784            ir_signature_to_abi_sig: FxHashMap::default(),
785            ir_sig_ref_to_abi_sig: SecondaryMap::with_capacity(func.dfg.signatures.len()),
786            abi_args: Vec::with_capacity(arg_estimate),
787            sigs: PrimaryMap::with_capacity(1 + func.dfg.signatures.len()),
788        };
789
790        sigs.make_abi_sig_from_ir_signature::<M>(func.signature.clone(), flags)?;
791        for sig_ref in func.dfg.signatures.keys() {
792            sigs.make_abi_sig_from_ir_sig_ref::<M>(sig_ref, &func.dfg, flags)?;
793        }
794
795        Ok(sigs)
796    }
797
798    /// Have we already interned an ABI signature for the given `ir::Signature`?
799    pub fn have_abi_sig_for_signature(&self, signature: &ir::Signature) -> bool {
800        self.ir_signature_to_abi_sig.contains_key(signature)
801    }
802
803    /// Construct and intern an ABI signature for the given `ir::Signature`.
804    pub fn make_abi_sig_from_ir_signature<M>(
805        &mut self,
806        signature: ir::Signature,
807        flags: &settings::Flags,
808    ) -> CodegenResult<Sig>
809    where
810        M: ABIMachineSpec,
811    {
812        // Because the `HashMap` entry API requires taking ownership of the
813        // lookup key -- and we want to avoid unnecessary clones of
814        // `ir::Signature`s, even at the cost of duplicate lookups -- we can't
815        // have a single, get-or-create-style method for interning
816        // `ir::Signature`s into ABI signatures. So at least (debug) assert that
817        // we aren't creating duplicate ABI signatures for the same
818        // `ir::Signature`.
819        debug_assert!(!self.have_abi_sig_for_signature(&signature));
820
821        let sig_data = self.from_func_sig::<M>(&signature, flags)?;
822        let sig = self.sigs.push(sig_data);
823        self.ir_signature_to_abi_sig.insert(signature, sig);
824        Ok(sig)
825    }
826
827    fn make_abi_sig_from_ir_sig_ref<M>(
828        &mut self,
829        sig_ref: ir::SigRef,
830        dfg: &ir::DataFlowGraph,
831        flags: &settings::Flags,
832    ) -> CodegenResult<Sig>
833    where
834        M: ABIMachineSpec,
835    {
836        if let Some(sig) = self.ir_sig_ref_to_abi_sig[sig_ref] {
837            return Ok(sig);
838        }
839        let signature = &dfg.signatures[sig_ref];
840        let sig_data = self.from_func_sig::<M>(signature, flags)?;
841        let sig = self.sigs.push(sig_data);
842        self.ir_sig_ref_to_abi_sig[sig_ref] = Some(sig);
843        Ok(sig)
844    }
845
846    /// Get the already-interned ABI signature id for the given `ir::SigRef`.
847    pub fn abi_sig_for_sig_ref(&self, sig_ref: ir::SigRef) -> Sig {
848        self.ir_sig_ref_to_abi_sig[sig_ref]
849            .expect("must call `make_abi_sig_from_ir_sig_ref` before `get_abi_sig_for_sig_ref`")
850    }
851
852    /// Get the already-interned ABI signature id for the given `ir::Signature`.
853    pub fn abi_sig_for_signature(&self, signature: &ir::Signature) -> Sig {
854        self.ir_signature_to_abi_sig
855            .get(signature)
856            .copied()
857            .expect("must call `make_abi_sig_from_ir_signature` before `get_abi_sig_for_signature`")
858    }
859
860    pub fn from_func_sig<M: ABIMachineSpec>(
861        &mut self,
862        sig: &ir::Signature,
863        flags: &settings::Flags,
864    ) -> CodegenResult<SigData> {
865        // Keep in sync with ensure_struct_return_ptr_is_returned
866        if sig.uses_special_return(ArgumentPurpose::StructReturn) {
867            panic!("Explicit StructReturn return value not allowed: {sig:?}")
868        }
869        let tmp;
870        let returns = if let Some(struct_ret_index) =
871            sig.special_param_index(ArgumentPurpose::StructReturn)
872        {
873            if !sig.returns.is_empty() {
874                panic!("No return values are allowed when using StructReturn: {sig:?}");
875            }
876            tmp = [sig.params[struct_ret_index]];
877            &tmp
878        } else {
879            sig.returns.as_slice()
880        };
881
882        // Compute args and retvals from signature. Handle retvals first,
883        // because we may need to add a return-area arg to the args.
884
885        // NOTE: We rely on the order of the args (rets -> args) inserted to compute the offsets in
886        // `SigSet::args()` and `SigSet::rets()`. Therefore, we cannot change the two
887        // compute_arg_locs order.
888        let (sized_stack_ret_space, _) = M::compute_arg_locs(
889            sig.call_conv,
890            flags,
891            &returns,
892            ArgsOrRets::Rets,
893            /* extra ret-area ptr = */ false,
894            ArgsAccumulator::new(&mut self.abi_args),
895        )?;
896        if !flags.enable_multi_ret_implicit_sret() {
897            assert_eq!(sized_stack_ret_space, 0);
898        }
899        let rets_end = u32::try_from(self.abi_args.len()).unwrap();
900
901        // To avoid overflow issues, limit the return size to something reasonable.
902        if sized_stack_ret_space > M::STACK_ARG_RET_SIZE_LIMIT {
903            return Err(CodegenError::ImplLimitExceeded);
904        }
905
906        let need_stack_return_area = sized_stack_ret_space > 0;
907        if need_stack_return_area {
908            assert!(!sig.uses_special_param(ir::ArgumentPurpose::StructReturn));
909        }
910
911        let (sized_stack_arg_space, stack_ret_arg) = M::compute_arg_locs(
912            sig.call_conv,
913            flags,
914            &sig.params,
915            ArgsOrRets::Args,
916            need_stack_return_area,
917            ArgsAccumulator::new(&mut self.abi_args),
918        )?;
919        let args_end = u32::try_from(self.abi_args.len()).unwrap();
920
921        // To avoid overflow issues, limit the arg size to something reasonable.
922        if sized_stack_arg_space > M::STACK_ARG_RET_SIZE_LIMIT {
923            return Err(CodegenError::ImplLimitExceeded);
924        }
925
926        trace!(
927            "ABISig: sig {:?} => args end = {} rets end = {}
928             arg stack = {} ret stack = {} stack_ret_arg = {:?}",
929            sig,
930            args_end,
931            rets_end,
932            sized_stack_arg_space,
933            sized_stack_ret_space,
934            need_stack_return_area,
935        );
936
937        let stack_ret_arg = stack_ret_arg.map(|s| u16::try_from(s).unwrap());
938        Ok(SigData {
939            args_end,
940            rets_end,
941            sized_stack_arg_space,
942            sized_stack_ret_space,
943            stack_ret_arg,
944            call_conv: sig.call_conv,
945        })
946    }
947
948    /// Get this signature's ABI arguments.
949    pub fn args(&self, sig: Sig) -> &[ABIArg] {
950        let sig_data = &self.sigs[sig];
951        // Please see comments in `SigSet::from_func_sig` of how we store the offsets.
952        let start = usize::try_from(sig_data.rets_end).unwrap();
953        let end = usize::try_from(sig_data.args_end).unwrap();
954        &self.abi_args[start..end]
955    }
956
957    /// Get information specifying how to pass the implicit pointer
958    /// to the return-value area on the stack, if required.
959    pub fn get_ret_arg(&self, sig: Sig) -> Option<ABIArg> {
960        let sig_data = &self.sigs[sig];
961        if let Some(i) = sig_data.stack_ret_arg {
962            Some(self.args(sig)[usize::from(i)].clone())
963        } else {
964            None
965        }
966    }
967
968    /// Get information specifying how to pass one argument.
969    pub fn get_arg(&self, sig: Sig, idx: usize) -> ABIArg {
970        self.args(sig)[idx].clone()
971    }
972
973    /// Get this signature's ABI returns.
974    pub fn rets(&self, sig: Sig) -> &[ABIArg] {
975        let sig_data = &self.sigs[sig];
976        // Please see comments in `SigSet::from_func_sig` of how we store the offsets.
977        let start = usize::try_from(sig.prev().map_or(0, |prev| self.sigs[prev].args_end)).unwrap();
978        let end = usize::try_from(sig_data.rets_end).unwrap();
979        &self.abi_args[start..end]
980    }
981
982    /// Get information specifying how to pass one return value.
983    pub fn get_ret(&self, sig: Sig, idx: usize) -> ABIArg {
984        self.rets(sig)[idx].clone()
985    }
986
987    /// Get the number of arguments expected.
988    pub fn num_args(&self, sig: Sig) -> usize {
989        let len = self.args(sig).len();
990        if self.sigs[sig].stack_ret_arg.is_some() {
991            len - 1
992        } else {
993            len
994        }
995    }
996
997    /// Get the number of return values expected.
998    pub fn num_rets(&self, sig: Sig) -> usize {
999        self.rets(sig).len()
1000    }
1001}
1002
1003// NB: we do _not_ implement `IndexMut` because these signatures are
1004// deduplicated and shared!
1005impl std::ops::Index<Sig> for SigSet {
1006    type Output = SigData;
1007
1008    fn index(&self, sig: Sig) -> &Self::Output {
1009        &self.sigs[sig]
1010    }
1011}
1012
1013/// Structure describing the layout of a function's stack frame.
1014#[derive(Clone, Debug, Default)]
1015pub struct FrameLayout {
1016    /// N.B. The areas whose sizes are given in this structure fully
1017    /// cover the current function's stack frame, from high to low
1018    /// stack addresses in the sequence below.  Each size contains
1019    /// any alignment padding that may be required by the ABI.
1020
1021    /// Size of incoming arguments on the stack.  This is not technically
1022    /// part of this function's frame, but code in the function will still
1023    /// need to access it.  Depending on the ABI, we may need to set up a
1024    /// frame pointer to do so; we also may need to pop this area from the
1025    /// stack upon return.
1026    pub incoming_args_size: u32,
1027
1028    /// The size of the incoming argument area, taking into account any
1029    /// potential increase in size required for tail calls present in the
1030    /// function. In the case that no tail calls are present, this value
1031    /// will be the same as [`Self::incoming_args_size`].
1032    pub tail_args_size: u32,
1033
1034    /// Size of the "setup area", typically holding the return address
1035    /// and/or the saved frame pointer.  This may be written either during
1036    /// the call itself (e.g. a pushed return address) or by code emitted
1037    /// from gen_prologue_frame_setup.  In any case, after that code has
1038    /// completed execution, the stack pointer is expected to point to the
1039    /// bottom of this area.  The same holds at the start of code emitted
1040    /// by gen_epilogue_frame_restore.
1041    pub setup_area_size: u32,
1042
1043    /// Size of the area used to save callee-saved clobbered registers.
1044    /// This area is accessed by code emitted from gen_clobber_save and
1045    /// gen_clobber_restore.
1046    pub clobber_size: u32,
1047
1048    /// Storage allocated for the fixed part of the stack frame.
1049    /// This contains stack slots and spill slots.
1050    pub fixed_frame_storage_size: u32,
1051
1052    /// The size of all stackslots.
1053    pub stackslots_size: u32,
1054
1055    /// Stack size to be reserved for outgoing arguments, if used by
1056    /// the current ABI, or 0 otherwise.  After gen_clobber_save and
1057    /// before gen_clobber_restore, the stack pointer points to the
1058    /// bottom of this area.
1059    pub outgoing_args_size: u32,
1060
1061    /// Sorted list of callee-saved registers that are clobbered
1062    /// according to the ABI.  These registers will be saved and
1063    /// restored by gen_clobber_save and gen_clobber_restore.
1064    pub clobbered_callee_saves: Vec<Writable<RealReg>>,
1065}
1066
1067impl FrameLayout {
1068    /// Split the clobbered callee-save registers into integer-class and
1069    /// float-class groups.
1070    ///
1071    /// This method does not currently support vector-class callee-save
1072    /// registers because no current backend has them.
1073    pub fn clobbered_callee_saves_by_class(&self) -> (&[Writable<RealReg>], &[Writable<RealReg>]) {
1074        let (ints, floats) = self.clobbered_callee_saves.split_at(
1075            self.clobbered_callee_saves
1076                .partition_point(|r| r.to_reg().class() == RegClass::Int),
1077        );
1078        debug_assert!(floats.iter().all(|r| r.to_reg().class() == RegClass::Float));
1079        (ints, floats)
1080    }
1081
1082    /// The size of FP to SP while the frame is active (not during prologue
1083    /// setup or epilogue tear down).
1084    pub fn active_size(&self) -> u32 {
1085        self.outgoing_args_size + self.fixed_frame_storage_size + self.clobber_size
1086    }
1087
1088    /// Get the offset from the SP to the sized stack slots area.
1089    pub fn sp_to_sized_stack_slots(&self) -> u32 {
1090        self.outgoing_args_size
1091    }
1092}
1093
1094/// ABI object for a function body.
1095pub struct Callee<M: ABIMachineSpec> {
1096    /// CLIF-level signature, possibly normalized.
1097    ir_sig: ir::Signature,
1098    /// Signature: arg and retval regs.
1099    sig: Sig,
1100    /// Defined dynamic types.
1101    dynamic_type_sizes: HashMap<Type, u32>,
1102    /// Offsets to each dynamic stackslot.
1103    dynamic_stackslots: PrimaryMap<DynamicStackSlot, u32>,
1104    /// Offsets to each sized stackslot.
1105    sized_stackslots: PrimaryMap<StackSlot, u32>,
1106    /// Total stack size of all stackslots
1107    stackslots_size: u32,
1108    /// Stack size to be reserved for outgoing arguments.
1109    outgoing_args_size: u32,
1110    /// Initially the number of bytes originating in the callers frame where stack arguments will
1111    /// live. After lowering this number may be larger than the size expected by the function being
1112    /// compiled, as tail calls potentially require more space for stack arguments.
1113    tail_args_size: u32,
1114    /// Register-argument defs, to be provided to the `args`
1115    /// pseudo-inst, and pregs to constrain them to.
1116    reg_args: Vec<ArgPair>,
1117    /// Finalized frame layout for this function.
1118    frame_layout: Option<FrameLayout>,
1119    /// The register holding the return-area pointer, if needed.
1120    ret_area_ptr: Option<Reg>,
1121    /// Calling convention this function expects.
1122    call_conv: isa::CallConv,
1123    /// The settings controlling this function's compilation.
1124    flags: settings::Flags,
1125    /// The ISA-specific flag values controlling this function's compilation.
1126    isa_flags: M::F,
1127    /// Whether or not this function is a "leaf", meaning it calls no other
1128    /// functions
1129    is_leaf: bool,
1130    /// If this function has a stack limit specified, then `Reg` is where the
1131    /// stack limit will be located after the instructions specified have been
1132    /// executed.
1133    ///
1134    /// Note that this is intended for insertion into the prologue, if
1135    /// present. Also note that because the instructions here execute in the
1136    /// prologue this happens after legalization/register allocation/etc so we
1137    /// need to be extremely careful with each instruction. The instructions are
1138    /// manually register-allocated and carefully only use caller-saved
1139    /// registers and keep nothing live after this sequence of instructions.
1140    stack_limit: Option<(Reg, SmallInstVec<M::I>)>,
1141
1142    _mach: PhantomData<M>,
1143}
1144
1145fn get_special_purpose_param_register(
1146    f: &ir::Function,
1147    sigs: &SigSet,
1148    sig: Sig,
1149    purpose: ir::ArgumentPurpose,
1150) -> Option<Reg> {
1151    let idx = f.signature.special_param_index(purpose)?;
1152    match &sigs.args(sig)[idx] {
1153        &ABIArg::Slots { ref slots, .. } => match &slots[0] {
1154            &ABIArgSlot::Reg { reg, .. } => Some(reg.into()),
1155            _ => None,
1156        },
1157        _ => None,
1158    }
1159}
1160
1161fn checked_round_up(val: u32, mask: u32) -> Option<u32> {
1162    Some(val.checked_add(mask)? & !mask)
1163}
1164
1165impl<M: ABIMachineSpec> Callee<M> {
1166    /// Create a new body ABI instance.
1167    pub fn new(
1168        f: &ir::Function,
1169        isa: &dyn TargetIsa,
1170        isa_flags: &M::F,
1171        sigs: &SigSet,
1172    ) -> CodegenResult<Self> {
1173        trace!("ABI: func signature {:?}", f.signature);
1174
1175        let flags = isa.flags().clone();
1176        let sig = sigs.abi_sig_for_signature(&f.signature);
1177
1178        let call_conv = f.signature.call_conv;
1179        // Only these calling conventions are supported.
1180        debug_assert!(
1181            call_conv == isa::CallConv::SystemV
1182                || call_conv == isa::CallConv::Tail
1183                || call_conv == isa::CallConv::Fast
1184                || call_conv == isa::CallConv::Cold
1185                || call_conv == isa::CallConv::WindowsFastcall
1186                || call_conv == isa::CallConv::AppleAarch64
1187                || call_conv == isa::CallConv::Winch,
1188            "Unsupported calling convention: {call_conv:?}"
1189        );
1190
1191        // Compute sized stackslot locations and total stackslot size.
1192        let mut end_offset: u32 = 0;
1193        let mut sized_stackslots = PrimaryMap::new();
1194
1195        for (stackslot, data) in f.sized_stack_slots.iter() {
1196            // We start our computation possibly unaligned where the previous
1197            // stackslot left off.
1198            let unaligned_start_offset = end_offset;
1199
1200            // The start of the stackslot must be aligned.
1201            //
1202            // We always at least machine-word-align slots, but also
1203            // satisfy the user's requested alignment.
1204            debug_assert!(data.align_shift < 32);
1205            let align = std::cmp::max(M::word_bytes(), 1u32 << data.align_shift);
1206            let mask = align - 1;
1207            let start_offset = checked_round_up(unaligned_start_offset, mask)
1208                .ok_or(CodegenError::ImplLimitExceeded)?;
1209
1210            // The end offset is the start offset increased by the size
1211            end_offset = start_offset
1212                .checked_add(data.size)
1213                .ok_or(CodegenError::ImplLimitExceeded)?;
1214
1215            debug_assert_eq!(stackslot.as_u32() as usize, sized_stackslots.len());
1216            sized_stackslots.push(start_offset);
1217        }
1218
1219        // Compute dynamic stackslot locations and total stackslot size.
1220        let mut dynamic_stackslots = PrimaryMap::new();
1221        for (stackslot, data) in f.dynamic_stack_slots.iter() {
1222            debug_assert_eq!(stackslot.as_u32() as usize, dynamic_stackslots.len());
1223
1224            // This computation is similar to the stackslots above
1225            let unaligned_start_offset = end_offset;
1226
1227            let mask = M::word_bytes() - 1;
1228            let start_offset = checked_round_up(unaligned_start_offset, mask)
1229                .ok_or(CodegenError::ImplLimitExceeded)?;
1230
1231            let ty = f.get_concrete_dynamic_ty(data.dyn_ty).ok_or_else(|| {
1232                CodegenError::Unsupported(format!("invalid dynamic vector type: {}", data.dyn_ty))
1233            })?;
1234
1235            end_offset = start_offset
1236                .checked_add(isa.dynamic_vector_bytes(ty))
1237                .ok_or(CodegenError::ImplLimitExceeded)?;
1238
1239            dynamic_stackslots.push(start_offset);
1240        }
1241
1242        // The size of the stackslots needs to be word aligned
1243        let stackslots_size = checked_round_up(end_offset, M::word_bytes() - 1)
1244            .ok_or(CodegenError::ImplLimitExceeded)?;
1245
1246        let mut dynamic_type_sizes = HashMap::with_capacity(f.dfg.dynamic_types.len());
1247        for (dyn_ty, _data) in f.dfg.dynamic_types.iter() {
1248            let ty = f
1249                .get_concrete_dynamic_ty(dyn_ty)
1250                .unwrap_or_else(|| panic!("invalid dynamic vector type: {dyn_ty}"));
1251            let size = isa.dynamic_vector_bytes(ty);
1252            dynamic_type_sizes.insert(ty, size);
1253        }
1254
1255        // Figure out what instructions, if any, will be needed to check the
1256        // stack limit. This can either be specified as a special-purpose
1257        // argument or as a global value which often calculates the stack limit
1258        // from the arguments.
1259        let stack_limit = f
1260            .stack_limit
1261            .map(|gv| gen_stack_limit::<M>(f, sigs, sig, gv));
1262
1263        let tail_args_size = sigs[sig].sized_stack_arg_space;
1264
1265        Ok(Self {
1266            ir_sig: ensure_struct_return_ptr_is_returned(&f.signature),
1267            sig,
1268            dynamic_stackslots,
1269            dynamic_type_sizes,
1270            sized_stackslots,
1271            stackslots_size,
1272            outgoing_args_size: 0,
1273            tail_args_size,
1274            reg_args: vec![],
1275            frame_layout: None,
1276            ret_area_ptr: None,
1277            call_conv,
1278            flags,
1279            isa_flags: isa_flags.clone(),
1280            is_leaf: f.is_leaf(),
1281            stack_limit,
1282            _mach: PhantomData,
1283        })
1284    }
1285
1286    /// Inserts instructions necessary for checking the stack limit into the
1287    /// prologue.
1288    ///
1289    /// This function will generate instructions necessary for perform a stack
1290    /// check at the header of a function. The stack check is intended to trap
1291    /// if the stack pointer goes below a particular threshold, preventing stack
1292    /// overflow in wasm or other code. The `stack_limit` argument here is the
1293    /// register which holds the threshold below which we're supposed to trap.
1294    /// This function is known to allocate `stack_size` bytes and we'll push
1295    /// instructions onto `insts`.
1296    ///
1297    /// Note that the instructions generated here are special because this is
1298    /// happening so late in the pipeline (e.g. after register allocation). This
1299    /// means that we need to do manual register allocation here and also be
1300    /// careful to not clobber any callee-saved or argument registers. For now
1301    /// this routine makes do with the `spilltmp_reg` as one temporary
1302    /// register, and a second register of `tmp2` which is caller-saved. This
1303    /// should be fine for us since no spills should happen in this sequence of
1304    /// instructions, so our register won't get accidentally clobbered.
1305    ///
1306    /// No values can be live after the prologue, but in this case that's ok
1307    /// because we just need to perform a stack check before progressing with
1308    /// the rest of the function.
1309    fn insert_stack_check(
1310        &self,
1311        stack_limit: Reg,
1312        stack_size: u32,
1313        insts: &mut SmallInstVec<M::I>,
1314    ) {
1315        // With no explicit stack allocated we can just emit the simple check of
1316        // the stack registers against the stack limit register, and trap if
1317        // it's out of bounds.
1318        if stack_size == 0 {
1319            insts.extend(M::gen_stack_lower_bound_trap(stack_limit));
1320            return;
1321        }
1322
1323        // Note that the 32k stack size here is pretty special. See the
1324        // documentation in x86/abi.rs for why this is here. The general idea is
1325        // that we're protecting against overflow in the addition that happens
1326        // below.
1327        if stack_size >= 32 * 1024 {
1328            insts.extend(M::gen_stack_lower_bound_trap(stack_limit));
1329        }
1330
1331        // Add the `stack_size` to `stack_limit`, placing the result in
1332        // `scratch`.
1333        //
1334        // Note though that `stack_limit`'s register may be the same as
1335        // `scratch`. If our stack size doesn't fit into an immediate this
1336        // means we need a second scratch register for loading the stack size
1337        // into a register.
1338        let scratch = Writable::from_reg(M::get_stacklimit_reg(self.call_conv));
1339        insts.extend(M::gen_add_imm(
1340            self.call_conv,
1341            scratch,
1342            stack_limit,
1343            stack_size,
1344        ));
1345        insts.extend(M::gen_stack_lower_bound_trap(scratch.to_reg()));
1346    }
1347}
1348
1349/// Generates the instructions necessary for the `gv` to be materialized into a
1350/// register.
1351///
1352/// This function will return a register that will contain the result of
1353/// evaluating `gv`. It will also return any instructions necessary to calculate
1354/// the value of the register.
1355///
1356/// Note that global values are typically lowered to instructions via the
1357/// standard legalization pass. Unfortunately though prologue generation happens
1358/// so late in the pipeline that we can't use these legalization passes to
1359/// generate the instructions for `gv`. As a result we duplicate some lowering
1360/// of `gv` here and support only some global values. This is similar to what
1361/// the x86 backend does for now, and hopefully this can be somewhat cleaned up
1362/// in the future too!
1363///
1364/// Also note that this function will make use of `writable_spilltmp_reg()` as a
1365/// temporary register to store values in if necessary. Currently after we write
1366/// to this register there's guaranteed to be no spilled values between where
1367/// it's used, because we're not participating in register allocation anyway!
1368fn gen_stack_limit<M: ABIMachineSpec>(
1369    f: &ir::Function,
1370    sigs: &SigSet,
1371    sig: Sig,
1372    gv: ir::GlobalValue,
1373) -> (Reg, SmallInstVec<M::I>) {
1374    let mut insts = smallvec![];
1375    let reg = generate_gv::<M>(f, sigs, sig, gv, &mut insts);
1376    return (reg, insts);
1377}
1378
1379fn generate_gv<M: ABIMachineSpec>(
1380    f: &ir::Function,
1381    sigs: &SigSet,
1382    sig: Sig,
1383    gv: ir::GlobalValue,
1384    insts: &mut SmallInstVec<M::I>,
1385) -> Reg {
1386    match f.global_values[gv] {
1387        // Return the direct register the vmcontext is in
1388        ir::GlobalValueData::VMContext => {
1389            get_special_purpose_param_register(f, sigs, sig, ir::ArgumentPurpose::VMContext)
1390                .expect("no vmcontext parameter found")
1391        }
1392        // Load our base value into a register, then load from that register
1393        // in to a temporary register.
1394        ir::GlobalValueData::Load {
1395            base,
1396            offset,
1397            global_type: _,
1398            flags: _,
1399        } => {
1400            let base = generate_gv::<M>(f, sigs, sig, base, insts);
1401            let into_reg = Writable::from_reg(M::get_stacklimit_reg(f.stencil.signature.call_conv));
1402            insts.push(M::gen_load_base_offset(
1403                into_reg,
1404                base,
1405                offset.into(),
1406                M::word_type(),
1407            ));
1408            return into_reg.to_reg();
1409        }
1410        ref other => panic!("global value for stack limit not supported: {other}"),
1411    }
1412}
1413
1414/// Returns true if the signature needs to be legalized.
1415fn missing_struct_return(sig: &ir::Signature) -> bool {
1416    sig.uses_special_param(ArgumentPurpose::StructReturn)
1417        && !sig.uses_special_return(ArgumentPurpose::StructReturn)
1418}
1419
1420fn ensure_struct_return_ptr_is_returned(sig: &ir::Signature) -> ir::Signature {
1421    // Keep in sync with Callee::new
1422    let mut sig = sig.clone();
1423    if sig.uses_special_return(ArgumentPurpose::StructReturn) {
1424        panic!("Explicit StructReturn return value not allowed: {sig:?}")
1425    }
1426    if let Some(struct_ret_index) = sig.special_param_index(ArgumentPurpose::StructReturn) {
1427        if !sig.returns.is_empty() {
1428            panic!("No return values are allowed when using StructReturn: {sig:?}");
1429        }
1430        sig.returns.insert(0, sig.params[struct_ret_index]);
1431    }
1432    sig
1433}
1434
1435/// ### Pre-Regalloc Functions
1436///
1437/// These methods of `Callee` may only be called before regalloc.
1438impl<M: ABIMachineSpec> Callee<M> {
1439    /// Access the (possibly legalized) signature.
1440    pub fn signature(&self) -> &ir::Signature {
1441        debug_assert!(
1442            !missing_struct_return(&self.ir_sig),
1443            "`Callee::ir_sig` is always legalized"
1444        );
1445        &self.ir_sig
1446    }
1447
1448    /// Initialize. This is called after the Callee is constructed because it
1449    /// may allocate a temp vreg, which can only be allocated once the lowering
1450    /// context exists.
1451    pub fn init_retval_area(
1452        &mut self,
1453        sigs: &SigSet,
1454        vregs: &mut VRegAllocator<M::I>,
1455    ) -> CodegenResult<()> {
1456        if sigs[self.sig].stack_ret_arg.is_some() {
1457            let ret_area_ptr = vregs.alloc(M::word_type())?;
1458            self.ret_area_ptr = Some(ret_area_ptr.only_reg().unwrap());
1459        }
1460        Ok(())
1461    }
1462
1463    /// Get the return area pointer register, if any.
1464    pub fn ret_area_ptr(&self) -> Option<Reg> {
1465        self.ret_area_ptr
1466    }
1467
1468    /// Accumulate outgoing arguments.
1469    ///
1470    /// This ensures that at least `size` bytes are allocated in the prologue to
1471    /// be available for use in function calls to hold arguments and/or return
1472    /// values. If this function is called multiple times, the maximum of all
1473    /// `size` values will be available.
1474    pub fn accumulate_outgoing_args_size(&mut self, size: u32) {
1475        if size > self.outgoing_args_size {
1476            self.outgoing_args_size = size;
1477        }
1478    }
1479
1480    /// Accumulate the incoming argument area size requirements for a tail call,
1481    /// as it could be larger than the incoming arguments of the function
1482    /// currently being compiled.
1483    pub fn accumulate_tail_args_size(&mut self, size: u32) {
1484        if size > self.tail_args_size {
1485            self.tail_args_size = size;
1486        }
1487    }
1488
1489    pub fn is_forward_edge_cfi_enabled(&self) -> bool {
1490        self.isa_flags.is_forward_edge_cfi_enabled()
1491    }
1492
1493    /// Get the calling convention implemented by this ABI object.
1494    pub fn call_conv(&self) -> isa::CallConv {
1495        self.call_conv
1496    }
1497
1498    /// Get the ABI-dependent MachineEnv for managing register allocation.
1499    pub fn machine_env(&self) -> &MachineEnv {
1500        M::get_machine_env(&self.flags, self.call_conv)
1501    }
1502
1503    /// The offsets of all sized stack slots (not spill slots) for debuginfo purposes.
1504    pub fn sized_stackslot_offsets(&self) -> &PrimaryMap<StackSlot, u32> {
1505        &self.sized_stackslots
1506    }
1507
1508    /// The offsets of all dynamic stack slots (not spill slots) for debuginfo purposes.
1509    pub fn dynamic_stackslot_offsets(&self) -> &PrimaryMap<DynamicStackSlot, u32> {
1510        &self.dynamic_stackslots
1511    }
1512
1513    /// Generate an instruction which copies an argument to a destination
1514    /// register.
1515    pub fn gen_copy_arg_to_regs(
1516        &mut self,
1517        sigs: &SigSet,
1518        idx: usize,
1519        into_regs: ValueRegs<Writable<Reg>>,
1520        vregs: &mut VRegAllocator<M::I>,
1521    ) -> SmallInstVec<M::I> {
1522        let mut insts = smallvec![];
1523        let mut copy_arg_slot_to_reg = |slot: &ABIArgSlot, into_reg: &Writable<Reg>| {
1524            match slot {
1525                &ABIArgSlot::Reg { reg, .. } => {
1526                    // Add a preg -> def pair to the eventual `args`
1527                    // instruction.  Extension mode doesn't matter
1528                    // (we're copying out, not in; we ignore high bits
1529                    // by convention).
1530                    let arg = ArgPair {
1531                        vreg: *into_reg,
1532                        preg: reg.into(),
1533                    };
1534                    self.reg_args.push(arg);
1535                }
1536                &ABIArgSlot::Stack {
1537                    offset,
1538                    ty,
1539                    extension,
1540                    ..
1541                } => {
1542                    // However, we have to respect the extension mode for stack
1543                    // slots, or else we grab the wrong bytes on big-endian.
1544                    let ext = M::get_ext_mode(sigs[self.sig].call_conv, extension);
1545                    let ty =
1546                        if ext != ArgumentExtension::None && M::word_bits() > ty_bits(ty) as u32 {
1547                            M::word_type()
1548                        } else {
1549                            ty
1550                        };
1551                    insts.push(M::gen_load_stack(
1552                        StackAMode::IncomingArg(offset, sigs[self.sig].sized_stack_arg_space),
1553                        *into_reg,
1554                        ty,
1555                    ));
1556                }
1557            }
1558        };
1559
1560        match &sigs.args(self.sig)[idx] {
1561            &ABIArg::Slots { ref slots, .. } => {
1562                assert_eq!(into_regs.len(), slots.len());
1563                for (slot, into_reg) in slots.iter().zip(into_regs.regs().iter()) {
1564                    copy_arg_slot_to_reg(&slot, &into_reg);
1565                }
1566            }
1567            &ABIArg::StructArg { offset, .. } => {
1568                let into_reg = into_regs.only_reg().unwrap();
1569                // Buffer address is implicitly defined by the ABI.
1570                insts.push(M::gen_get_stack_addr(
1571                    StackAMode::IncomingArg(offset, sigs[self.sig].sized_stack_arg_space),
1572                    into_reg,
1573                ));
1574            }
1575            &ABIArg::ImplicitPtrArg { pointer, ty, .. } => {
1576                let into_reg = into_regs.only_reg().unwrap();
1577                // We need to dereference the pointer.
1578                let base = match &pointer {
1579                    &ABIArgSlot::Reg { reg, ty, .. } => {
1580                        let tmp = vregs.alloc_with_deferred_error(ty).only_reg().unwrap();
1581                        self.reg_args.push(ArgPair {
1582                            vreg: Writable::from_reg(tmp),
1583                            preg: reg.into(),
1584                        });
1585                        tmp
1586                    }
1587                    &ABIArgSlot::Stack { offset, ty, .. } => {
1588                        let addr_reg = writable_value_regs(vregs.alloc_with_deferred_error(ty))
1589                            .only_reg()
1590                            .unwrap();
1591                        insts.push(M::gen_load_stack(
1592                            StackAMode::IncomingArg(offset, sigs[self.sig].sized_stack_arg_space),
1593                            addr_reg,
1594                            ty,
1595                        ));
1596                        addr_reg.to_reg()
1597                    }
1598                };
1599                insts.push(M::gen_load_base_offset(into_reg, base, 0, ty));
1600            }
1601        }
1602        insts
1603    }
1604
1605    /// Generate an instruction which copies a source register to a return value slot.
1606    pub fn gen_copy_regs_to_retval(
1607        &self,
1608        sigs: &SigSet,
1609        idx: usize,
1610        from_regs: ValueRegs<Reg>,
1611        vregs: &mut VRegAllocator<M::I>,
1612    ) -> (SmallVec<[RetPair; 2]>, SmallInstVec<M::I>) {
1613        let mut reg_pairs = smallvec![];
1614        let mut ret = smallvec![];
1615        let word_bits = M::word_bits() as u8;
1616        match &sigs.rets(self.sig)[idx] {
1617            &ABIArg::Slots { ref slots, .. } => {
1618                assert_eq!(from_regs.len(), slots.len());
1619                for (slot, &from_reg) in slots.iter().zip(from_regs.regs().iter()) {
1620                    match slot {
1621                        &ABIArgSlot::Reg {
1622                            reg, ty, extension, ..
1623                        } => {
1624                            let from_bits = ty_bits(ty) as u8;
1625                            let ext = M::get_ext_mode(sigs[self.sig].call_conv, extension);
1626                            let vreg = match (ext, from_bits) {
1627                                (ir::ArgumentExtension::Uext, n)
1628                                | (ir::ArgumentExtension::Sext, n)
1629                                    if n < word_bits =>
1630                                {
1631                                    let signed = ext == ir::ArgumentExtension::Sext;
1632                                    let dst =
1633                                        writable_value_regs(vregs.alloc_with_deferred_error(ty))
1634                                            .only_reg()
1635                                            .unwrap();
1636                                    ret.push(M::gen_extend(
1637                                        dst, from_reg, signed, from_bits,
1638                                        /* to_bits = */ word_bits,
1639                                    ));
1640                                    dst.to_reg()
1641                                }
1642                                _ => {
1643                                    // No move needed, regalloc2 will emit it using the constraint
1644                                    // added by the RetPair.
1645                                    from_reg
1646                                }
1647                            };
1648                            reg_pairs.push(RetPair {
1649                                vreg,
1650                                preg: Reg::from(reg),
1651                            });
1652                        }
1653                        &ABIArgSlot::Stack {
1654                            offset,
1655                            ty,
1656                            extension,
1657                            ..
1658                        } => {
1659                            let mut ty = ty;
1660                            let from_bits = ty_bits(ty) as u8;
1661                            // A machine ABI implementation should ensure that stack frames
1662                            // have "reasonable" size. All current ABIs for machinst
1663                            // backends (aarch64 and x64) enforce a 128MB limit.
1664                            let off = i32::try_from(offset).expect(
1665                                "Argument stack offset greater than 2GB; should hit impl limit first",
1666                                );
1667                            let ext = M::get_ext_mode(sigs[self.sig].call_conv, extension);
1668                            // Trash the from_reg; it should be its last use.
1669                            match (ext, from_bits) {
1670                                (ir::ArgumentExtension::Uext, n)
1671                                | (ir::ArgumentExtension::Sext, n)
1672                                    if n < word_bits =>
1673                                {
1674                                    assert_eq!(M::word_reg_class(), from_reg.class());
1675                                    let signed = ext == ir::ArgumentExtension::Sext;
1676                                    let dst =
1677                                        writable_value_regs(vregs.alloc_with_deferred_error(ty))
1678                                            .only_reg()
1679                                            .unwrap();
1680                                    ret.push(M::gen_extend(
1681                                        dst, from_reg, signed, from_bits,
1682                                        /* to_bits = */ word_bits,
1683                                    ));
1684                                    // Store the extended version.
1685                                    ty = M::word_type();
1686                                }
1687                                _ => {}
1688                            };
1689                            ret.push(M::gen_store_base_offset(
1690                                self.ret_area_ptr.unwrap(),
1691                                off,
1692                                from_reg,
1693                                ty,
1694                            ));
1695                        }
1696                    }
1697                }
1698            }
1699            ABIArg::StructArg { .. } => {
1700                panic!("StructArg in return position is unsupported");
1701            }
1702            ABIArg::ImplicitPtrArg { .. } => {
1703                panic!("ImplicitPtrArg in return position is unsupported");
1704            }
1705        }
1706        (reg_pairs, ret)
1707    }
1708
1709    /// Generate any setup instruction needed to save values to the
1710    /// return-value area. This is usually used when were are multiple return
1711    /// values or an otherwise large return value that must be passed on the
1712    /// stack; typically the ABI specifies an extra hidden argument that is a
1713    /// pointer to that memory.
1714    pub fn gen_retval_area_setup(
1715        &mut self,
1716        sigs: &SigSet,
1717        vregs: &mut VRegAllocator<M::I>,
1718    ) -> Option<M::I> {
1719        if let Some(i) = sigs[self.sig].stack_ret_arg {
1720            let ret_area_ptr = Writable::from_reg(self.ret_area_ptr.unwrap());
1721            let insts =
1722                self.gen_copy_arg_to_regs(sigs, i.into(), ValueRegs::one(ret_area_ptr), vregs);
1723            insts.into_iter().next().map(|inst| {
1724                trace!(
1725                    "gen_retval_area_setup: inst {:?}; ptr reg is {:?}",
1726                    inst,
1727                    ret_area_ptr.to_reg()
1728                );
1729                inst
1730            })
1731        } else {
1732            trace!("gen_retval_area_setup: not needed");
1733            None
1734        }
1735    }
1736
1737    /// Generate a return instruction.
1738    pub fn gen_rets(&self, rets: Vec<RetPair>) -> M::I {
1739        M::gen_rets(rets)
1740    }
1741
1742    /// Set up arguments values `args` for a call with signature `sig`.
1743    /// This will return a series of instructions to be emitted to set
1744    /// up all arguments, as well as a `CallArgList` list representing
1745    /// the arguments passed in registers.  The latter need to be added
1746    /// as constraints to the actual call instruction.
1747    pub fn gen_call_args(
1748        &self,
1749        sigs: &SigSet,
1750        sig: Sig,
1751        args: &[ValueRegs<Reg>],
1752        is_tail_call: bool,
1753        flags: &settings::Flags,
1754        vregs: &mut VRegAllocator<M::I>,
1755    ) -> (CallArgList, SmallInstVec<M::I>) {
1756        let mut uses: CallArgList = smallvec![];
1757        let mut insts = smallvec![];
1758
1759        assert_eq!(args.len(), sigs.num_args(sig));
1760
1761        let call_conv = sigs[sig].call_conv;
1762        let stack_arg_space = sigs[sig].sized_stack_arg_space;
1763        let stack_arg = |offset| {
1764            if is_tail_call {
1765                StackAMode::IncomingArg(offset, stack_arg_space)
1766            } else {
1767                StackAMode::OutgoingArg(offset)
1768            }
1769        };
1770
1771        let word_ty = M::word_type();
1772        let word_rc = M::word_reg_class();
1773        let word_bits = M::word_bits() as usize;
1774
1775        if is_tail_call {
1776            debug_assert_eq!(
1777                self.call_conv,
1778                isa::CallConv::Tail,
1779                "Can only do `return_call`s from within a `tail` calling convention function"
1780            );
1781        }
1782
1783        // Helper to process a single argument slot (register or stack slot).
1784        // This will either add the register to the `uses` list or write the
1785        // value to the stack slot in the outgoing argument area (or for tail
1786        // calls, the incoming argument area).
1787        let mut process_arg_slot = |insts: &mut SmallInstVec<M::I>, slot, vreg, ty| {
1788            match &slot {
1789                &ABIArgSlot::Reg { reg, .. } => {
1790                    uses.push(CallArgPair {
1791                        vreg,
1792                        preg: reg.into(),
1793                    });
1794                }
1795                &ABIArgSlot::Stack { offset, .. } => {
1796                    insts.push(M::gen_store_stack(stack_arg(offset), vreg, ty));
1797                }
1798            };
1799        };
1800
1801        // First pass: Handle `StructArg` arguments.  These need to be copied
1802        // into their associated stack buffers.  This should happen before any
1803        // of the other arguments are processed, as the `memcpy` call might
1804        // clobber registers used by other arguments.
1805        for (idx, from_regs) in args.iter().enumerate() {
1806            match &sigs.args(sig)[idx] {
1807                &ABIArg::Slots { .. } | &ABIArg::ImplicitPtrArg { .. } => {}
1808                &ABIArg::StructArg { offset, size, .. } => {
1809                    let tmp = vregs.alloc_with_deferred_error(word_ty).only_reg().unwrap();
1810                    insts.push(M::gen_get_stack_addr(
1811                        stack_arg(offset),
1812                        Writable::from_reg(tmp),
1813                    ));
1814                    insts.extend(M::gen_memcpy(
1815                        isa::CallConv::for_libcall(flags, call_conv),
1816                        tmp,
1817                        from_regs.only_reg().unwrap(),
1818                        size as usize,
1819                        |ty| {
1820                            Writable::from_reg(
1821                                vregs.alloc_with_deferred_error(ty).only_reg().unwrap(),
1822                            )
1823                        },
1824                    ));
1825                }
1826            }
1827        }
1828
1829        // Second pass: Handle everything except `StructArg` arguments.
1830        for (idx, from_regs) in args.iter().enumerate() {
1831            match sigs.args(sig)[idx] {
1832                ABIArg::Slots { ref slots, .. } => {
1833                    assert_eq!(from_regs.len(), slots.len());
1834                    for (slot, from_reg) in slots.iter().zip(from_regs.regs().iter()) {
1835                        // Load argument slot value from `from_reg`, and perform any zero-
1836                        // or sign-extension that is required by the ABI.
1837                        let (ty, extension) = match *slot {
1838                            ABIArgSlot::Reg { ty, extension, .. } => (ty, extension),
1839                            ABIArgSlot::Stack { ty, extension, .. } => (ty, extension),
1840                        };
1841                        let ext = M::get_ext_mode(call_conv, extension);
1842                        let (vreg, ty) = if ext != ir::ArgumentExtension::None
1843                            && ty_bits(ty) < word_bits
1844                        {
1845                            assert_eq!(word_rc, from_reg.class());
1846                            let signed = match ext {
1847                                ir::ArgumentExtension::Uext => false,
1848                                ir::ArgumentExtension::Sext => true,
1849                                _ => unreachable!(),
1850                            };
1851                            let tmp = vregs.alloc_with_deferred_error(word_ty).only_reg().unwrap();
1852                            insts.push(M::gen_extend(
1853                                Writable::from_reg(tmp),
1854                                *from_reg,
1855                                signed,
1856                                ty_bits(ty) as u8,
1857                                word_bits as u8,
1858                            ));
1859                            (tmp, word_ty)
1860                        } else {
1861                            (*from_reg, ty)
1862                        };
1863                        process_arg_slot(&mut insts, *slot, vreg, ty);
1864                    }
1865                }
1866                ABIArg::ImplicitPtrArg {
1867                    offset,
1868                    pointer,
1869                    ty,
1870                    ..
1871                } => {
1872                    let vreg = from_regs.only_reg().unwrap();
1873                    let tmp = vregs.alloc_with_deferred_error(word_ty).only_reg().unwrap();
1874                    insts.push(M::gen_get_stack_addr(
1875                        stack_arg(offset),
1876                        Writable::from_reg(tmp),
1877                    ));
1878                    insts.push(M::gen_store_base_offset(tmp, 0, vreg, ty));
1879                    process_arg_slot(&mut insts, pointer, tmp, word_ty);
1880                }
1881                ABIArg::StructArg { .. } => {}
1882            }
1883        }
1884
1885        // Finally, set the stack-return pointer to the return argument area.
1886        // For tail calls, this means forwarding the incoming stack-return pointer.
1887        if let Some(ret_arg) = sigs.get_ret_arg(sig) {
1888            let ret_area = if is_tail_call {
1889                self.ret_area_ptr.expect(
1890                    "if the tail callee has a return pointer, then the tail caller must as well",
1891                )
1892            } else {
1893                let tmp = vregs.alloc_with_deferred_error(word_ty).only_reg().unwrap();
1894                let amode = StackAMode::OutgoingArg(stack_arg_space.into());
1895                insts.push(M::gen_get_stack_addr(amode, Writable::from_reg(tmp)));
1896                tmp
1897            };
1898            match ret_arg {
1899                // The return pointer must occupy a single slot.
1900                ABIArg::Slots { slots, .. } => {
1901                    assert_eq!(slots.len(), 1);
1902                    process_arg_slot(&mut insts, slots[0], ret_area, word_ty);
1903                }
1904                _ => unreachable!(),
1905            }
1906        }
1907
1908        (uses, insts)
1909    }
1910
1911    /// Set up return values `outputs` for a call with signature `sig`.
1912    /// This does not emit (or return) any instructions, but returns a
1913    /// `CallRetList` representing the return value constraints.  This
1914    /// needs to be added to the actual call instruction.
1915    ///
1916    /// If `try_call_payloads` is non-zero, it is expected to hold
1917    /// exception payload registers for try_call instructions.  These
1918    /// will be added as needed to the `CallRetList` as well.
1919    pub fn gen_call_rets(
1920        &self,
1921        sigs: &SigSet,
1922        sig: Sig,
1923        outputs: &[ValueRegs<Reg>],
1924        try_call_payloads: Option<&[Writable<Reg>]>,
1925        vregs: &mut VRegAllocator<M::I>,
1926    ) -> CallRetList {
1927        let callee_conv = sigs[sig].call_conv;
1928        let stack_arg_space = sigs[sig].sized_stack_arg_space;
1929
1930        let word_ty = M::word_type();
1931        let word_bits = M::word_bits() as usize;
1932
1933        let mut defs: CallRetList = smallvec![];
1934        let mut outputs = outputs.into_iter();
1935        let num_rets = sigs.num_rets(sig);
1936        for idx in 0..num_rets {
1937            let ret = sigs.rets(sig)[idx].clone();
1938            match ret {
1939                ABIArg::Slots {
1940                    ref slots, purpose, ..
1941                } => {
1942                    // We do not use the returned copy of the return buffer pointer,
1943                    // so skip any StructReturn returns that may be present.
1944                    if purpose == ArgumentPurpose::StructReturn {
1945                        continue;
1946                    }
1947                    let retval_regs = outputs.next().unwrap();
1948                    assert_eq!(retval_regs.len(), slots.len());
1949                    for (slot, retval_reg) in slots.iter().zip(retval_regs.regs().iter()) {
1950                        // We do not perform any extension because we're copying out, not in,
1951                        // and we ignore high bits in our own registers by convention.  However,
1952                        // we still need to use the proper extended type to access stack slots
1953                        // (this is critical on big-endian systems).
1954                        let (ty, extension) = match *slot {
1955                            ABIArgSlot::Reg { ty, extension, .. } => (ty, extension),
1956                            ABIArgSlot::Stack { ty, extension, .. } => (ty, extension),
1957                        };
1958                        let ext = M::get_ext_mode(callee_conv, extension);
1959                        let ty = if ext != ir::ArgumentExtension::None && ty_bits(ty) < word_bits {
1960                            word_ty
1961                        } else {
1962                            ty
1963                        };
1964
1965                        match slot {
1966                            &ABIArgSlot::Reg { reg, .. } => {
1967                                defs.push(CallRetPair {
1968                                    vreg: Writable::from_reg(*retval_reg),
1969                                    location: RetLocation::Reg(reg.into(), ty),
1970                                });
1971                            }
1972                            &ABIArgSlot::Stack { offset, .. } => {
1973                                let amode =
1974                                    StackAMode::OutgoingArg(offset + i64::from(stack_arg_space));
1975                                defs.push(CallRetPair {
1976                                    vreg: Writable::from_reg(*retval_reg),
1977                                    location: RetLocation::Stack(amode, ty),
1978                                });
1979                            }
1980                        }
1981                    }
1982                }
1983                ABIArg::StructArg { .. } => {
1984                    panic!("StructArg not supported in return position");
1985                }
1986                ABIArg::ImplicitPtrArg { .. } => {
1987                    panic!("ImplicitPtrArg not supported in return position");
1988                }
1989            }
1990        }
1991        assert!(outputs.next().is_none());
1992
1993        if let Some(try_call_payloads) = try_call_payloads {
1994            // We need to update `defs` to contain the exception
1995            // payload regs as well. We have two sources of info that
1996            // we join:
1997            //
1998            // - The machine-specific ABI implementation `M`, which
1999            //   tells us the particular registers that payload values
2000            //   must be in
2001            // - The passed-in lowering context, which gives us the
2002            //   vregs we must define.
2003            //
2004            // Note that payload values may need to end up in the same
2005            // physical registers as ordinary return values; this is
2006            // not a conflict, because we either get one or the
2007            // other. For regalloc's purposes, we define both starting
2008            // here at the callsite, but we can share one def in the
2009            // `defs` list and alias one vreg to another. Thus we
2010            // handle the two cases below for each payload register:
2011            // overlaps a return value (and we alias to it) or not
2012            // (and we add a def).
2013            let pregs = M::exception_payload_regs(callee_conv);
2014            for (i, &preg) in pregs.iter().enumerate() {
2015                let vreg = try_call_payloads[i];
2016                if let Some(existing) = defs.iter().find(|def| match def.location {
2017                    RetLocation::Reg(r, _) => r == preg,
2018                    _ => false,
2019                }) {
2020                    vregs.set_vreg_alias(vreg.to_reg(), existing.vreg.to_reg());
2021                } else {
2022                    defs.push(CallRetPair {
2023                        vreg,
2024                        location: RetLocation::Reg(preg, word_ty),
2025                    });
2026                }
2027            }
2028        }
2029
2030        defs
2031    }
2032
2033    /// Populate a `CallInfo` for a call with signature `sig`.
2034    ///
2035    /// `dest` is the target-specific call destination value
2036    /// `uses` is the `CallArgList` describing argument constraints
2037    /// `defs` is the `CallRetList` describing return constraints
2038    /// `try_call_info` describes exception targets for try_call instructions
2039    ///
2040    /// The clobber list is computed here from the above data.
2041    pub fn gen_call_info<T>(
2042        &self,
2043        sigs: &SigSet,
2044        sig: Sig,
2045        dest: T,
2046        uses: CallArgList,
2047        defs: CallRetList,
2048        try_call_info: Option<TryCallInfo>,
2049    ) -> CallInfo<T> {
2050        let caller_conv = self.call_conv;
2051        let callee_conv = sigs[sig].call_conv;
2052        let stack_arg_space = sigs[sig].sized_stack_arg_space;
2053
2054        let clobbers = {
2055            // Get clobbers: all caller-saves. These may include return value
2056            // regs, which we will remove from the clobber set below.
2057            let mut clobbers =
2058                <M>::get_regs_clobbered_by_call(callee_conv, try_call_info.is_some());
2059
2060            // Remove retval regs from clobbers.
2061            for def in &defs {
2062                if let RetLocation::Reg(preg, _) = def.location {
2063                    clobbers.remove(PReg::from(preg.to_real_reg().unwrap()));
2064                }
2065            }
2066
2067            clobbers
2068        };
2069
2070        // Any adjustment to SP to account for required outgoing arguments/stack return values must
2071        // be done inside of the call pseudo-op, to ensure that SP is always in a consistent
2072        // state for all other instructions. For example, if a tail-call abi function is called
2073        // here, the reclamation of the outgoing argument area must be done inside of the call
2074        // pseudo-op's emission to ensure that SP is consistent at all other points in the lowered
2075        // function. (Except the prologue and epilogue, but those are fairly special parts of the
2076        // function that establish the SP invariants that are relied on elsewhere and are generated
2077        // after the register allocator has run and thus cannot have register allocator-inserted
2078        // references to SP offsets.)
2079
2080        let callee_pop_size = if callee_conv == isa::CallConv::Tail {
2081            // The tail calling convention has callees pop stack arguments.
2082            stack_arg_space
2083        } else {
2084            0
2085        };
2086
2087        CallInfo {
2088            dest,
2089            uses,
2090            defs,
2091            clobbers,
2092            callee_conv,
2093            caller_conv,
2094            callee_pop_size,
2095            try_call_info,
2096        }
2097    }
2098
2099    /// Produce an instruction that computes a sized stackslot address.
2100    pub fn sized_stackslot_addr(
2101        &self,
2102        slot: StackSlot,
2103        offset: u32,
2104        into_reg: Writable<Reg>,
2105    ) -> M::I {
2106        // Offset from beginning of stackslot area.
2107        let stack_off = self.sized_stackslots[slot] as i64;
2108        let sp_off: i64 = stack_off + (offset as i64);
2109        M::gen_get_stack_addr(StackAMode::Slot(sp_off), into_reg)
2110    }
2111
2112    /// Produce an instruction that computes a dynamic stackslot address.
2113    pub fn dynamic_stackslot_addr(&self, slot: DynamicStackSlot, into_reg: Writable<Reg>) -> M::I {
2114        let stack_off = self.dynamic_stackslots[slot] as i64;
2115        M::gen_get_stack_addr(StackAMode::Slot(stack_off), into_reg)
2116    }
2117
2118    /// Get an `args` pseudo-inst, if any, that should appear at the
2119    /// very top of the function body prior to regalloc.
2120    pub fn take_args(&mut self) -> Option<M::I> {
2121        if self.reg_args.len() > 0 {
2122            // Very first instruction is an `args` pseudo-inst that
2123            // establishes live-ranges for in-register arguments and
2124            // constrains them at the start of the function to the
2125            // locations defined by the ABI.
2126            Some(M::gen_args(std::mem::take(&mut self.reg_args)))
2127        } else {
2128            None
2129        }
2130    }
2131}
2132
2133/// ### Post-Regalloc Functions
2134///
2135/// These methods of `Callee` may only be called after
2136/// regalloc.
2137impl<M: ABIMachineSpec> Callee<M> {
2138    /// Compute the final frame layout, post-regalloc.
2139    ///
2140    /// This must be called before gen_prologue or gen_epilogue.
2141    pub fn compute_frame_layout(
2142        &mut self,
2143        sigs: &SigSet,
2144        spillslots: usize,
2145        clobbered: Vec<Writable<RealReg>>,
2146    ) {
2147        let bytes = M::word_bytes();
2148        let total_stacksize = self.stackslots_size + bytes * spillslots as u32;
2149        let mask = M::stack_align(self.call_conv) - 1;
2150        let total_stacksize = (total_stacksize + mask) & !mask; // 16-align the stack.
2151        self.frame_layout = Some(M::compute_frame_layout(
2152            self.call_conv,
2153            &self.flags,
2154            self.signature(),
2155            &clobbered,
2156            self.is_leaf,
2157            self.stack_args_size(sigs),
2158            self.tail_args_size,
2159            self.stackslots_size,
2160            total_stacksize,
2161            self.outgoing_args_size,
2162        ));
2163    }
2164
2165    /// Generate a prologue, post-regalloc.
2166    ///
2167    /// This should include any stack frame or other setup necessary to use the
2168    /// other methods (`load_arg`, `store_retval`, and spillslot accesses.)
2169    pub fn gen_prologue(&self) -> SmallInstVec<M::I> {
2170        let frame_layout = self.frame_layout();
2171        let mut insts = smallvec![];
2172
2173        // Set up frame.
2174        insts.extend(M::gen_prologue_frame_setup(
2175            self.call_conv,
2176            &self.flags,
2177            &self.isa_flags,
2178            &frame_layout,
2179        ));
2180
2181        // The stack limit check needs to cover all the stack adjustments we
2182        // might make, up to the next stack limit check in any function we
2183        // call. Since this happens after frame setup, the current function's
2184        // setup area needs to be accounted for in the caller's stack limit
2185        // check, but we need to account for any setup area that our callees
2186        // might need. Note that s390x may also use the outgoing args area for
2187        // backtrace support even in leaf functions, so that should be accounted
2188        // for unconditionally.
2189        let total_stacksize = (frame_layout.tail_args_size - frame_layout.incoming_args_size)
2190            + frame_layout.clobber_size
2191            + frame_layout.fixed_frame_storage_size
2192            + frame_layout.outgoing_args_size
2193            + if self.is_leaf {
2194                0
2195            } else {
2196                frame_layout.setup_area_size
2197            };
2198
2199        // Leaf functions with zero stack don't need a stack check if one's
2200        // specified, otherwise always insert the stack check.
2201        if total_stacksize > 0 || !self.is_leaf {
2202            if let Some((reg, stack_limit_load)) = &self.stack_limit {
2203                insts.extend(stack_limit_load.clone());
2204                self.insert_stack_check(*reg, total_stacksize, &mut insts);
2205            }
2206
2207            if self.flags.enable_probestack() {
2208                let guard_size = 1 << self.flags.probestack_size_log2();
2209                match self.flags.probestack_strategy() {
2210                    ProbestackStrategy::Inline => M::gen_inline_probestack(
2211                        &mut insts,
2212                        self.call_conv,
2213                        total_stacksize,
2214                        guard_size,
2215                    ),
2216                    ProbestackStrategy::Outline => {
2217                        if total_stacksize >= guard_size {
2218                            M::gen_probestack(&mut insts, total_stacksize);
2219                        }
2220                    }
2221                }
2222            }
2223        }
2224
2225        // Save clobbered registers.
2226        insts.extend(M::gen_clobber_save(
2227            self.call_conv,
2228            &self.flags,
2229            &frame_layout,
2230        ));
2231
2232        insts
2233    }
2234
2235    /// Generate an epilogue, post-regalloc.
2236    ///
2237    /// Note that this must generate the actual return instruction (rather than
2238    /// emitting this in the lowering logic), because the epilogue code comes
2239    /// before the return and the two are likely closely related.
2240    pub fn gen_epilogue(&self) -> SmallInstVec<M::I> {
2241        let frame_layout = self.frame_layout();
2242        let mut insts = smallvec![];
2243
2244        // Restore clobbered registers.
2245        insts.extend(M::gen_clobber_restore(
2246            self.call_conv,
2247            &self.flags,
2248            &frame_layout,
2249        ));
2250
2251        // Tear down frame.
2252        insts.extend(M::gen_epilogue_frame_restore(
2253            self.call_conv,
2254            &self.flags,
2255            &self.isa_flags,
2256            &frame_layout,
2257        ));
2258
2259        // And return.
2260        insts.extend(M::gen_return(
2261            self.call_conv,
2262            &self.isa_flags,
2263            &frame_layout,
2264        ));
2265
2266        trace!("Epilogue: {:?}", insts);
2267        insts
2268    }
2269
2270    /// Return a reference to the computed frame layout information. This
2271    /// function will panic if it's called before [`Self::compute_frame_layout`].
2272    pub fn frame_layout(&self) -> &FrameLayout {
2273        self.frame_layout
2274            .as_ref()
2275            .expect("frame layout not computed before prologue generation")
2276    }
2277
2278    /// Returns the full frame size for the given function, after prologue
2279    /// emission has run. This comprises the spill slots and stack-storage
2280    /// slots as well as storage for clobbered callee-save registers, but
2281    /// not arguments arguments pushed at callsites within this function,
2282    /// or other ephemeral pushes.
2283    pub fn frame_size(&self) -> u32 {
2284        let frame_layout = self.frame_layout();
2285        frame_layout.clobber_size + frame_layout.fixed_frame_storage_size
2286    }
2287
2288    /// Returns offset from the slot base in the current frame to the caller's SP.
2289    pub fn slot_base_to_caller_sp_offset(&self) -> u32 {
2290        let frame_layout = self.frame_layout();
2291        frame_layout.clobber_size
2292            + frame_layout.fixed_frame_storage_size
2293            + frame_layout.setup_area_size
2294    }
2295
2296    /// Returns the size of arguments expected on the stack.
2297    pub fn stack_args_size(&self, sigs: &SigSet) -> u32 {
2298        sigs[self.sig].sized_stack_arg_space
2299    }
2300
2301    /// Get the spill-slot size.
2302    pub fn get_spillslot_size(&self, rc: RegClass) -> u32 {
2303        let max = if self.dynamic_type_sizes.len() == 0 {
2304            16
2305        } else {
2306            *self
2307                .dynamic_type_sizes
2308                .iter()
2309                .max_by(|x, y| x.1.cmp(&y.1))
2310                .map(|(_k, v)| v)
2311                .unwrap()
2312        };
2313        M::get_number_of_spillslots_for_value(rc, max, &self.isa_flags)
2314    }
2315
2316    /// Get the spill slot offset relative to the fixed allocation area start.
2317    pub fn get_spillslot_offset(&self, slot: SpillSlot) -> i64 {
2318        // Offset from beginning of spillslot area.
2319        let islot = slot.index() as i64;
2320        let spill_off = islot * M::word_bytes() as i64;
2321        let sp_off = self.stackslots_size as i64 + spill_off;
2322
2323        sp_off
2324    }
2325
2326    /// Generate a spill.
2327    pub fn gen_spill(&self, to_slot: SpillSlot, from_reg: RealReg) -> M::I {
2328        let ty = M::I::canonical_type_for_rc(from_reg.class());
2329        debug_assert_eq!(<M>::I::rc_for_type(ty).unwrap().1, &[ty]);
2330
2331        let sp_off = self.get_spillslot_offset(to_slot);
2332        trace!("gen_spill: {from_reg:?} into slot {to_slot:?} at offset {sp_off}");
2333
2334        let from = StackAMode::Slot(sp_off);
2335        <M>::gen_store_stack(from, Reg::from(from_reg), ty)
2336    }
2337
2338    /// Generate a reload (fill).
2339    pub fn gen_reload(&self, to_reg: Writable<RealReg>, from_slot: SpillSlot) -> M::I {
2340        let ty = M::I::canonical_type_for_rc(to_reg.to_reg().class());
2341        debug_assert_eq!(<M>::I::rc_for_type(ty).unwrap().1, &[ty]);
2342
2343        let sp_off = self.get_spillslot_offset(from_slot);
2344        trace!("gen_reload: {to_reg:?} from slot {from_slot:?} at offset {sp_off}");
2345
2346        let from = StackAMode::Slot(sp_off);
2347        <M>::gen_load_stack(from, to_reg.map(Reg::from), ty)
2348    }
2349}
2350
2351/// An input argument to a call instruction: the vreg that is used,
2352/// and the preg it is constrained to (per the ABI).
2353#[derive(Clone, Debug)]
2354pub struct CallArgPair {
2355    /// The virtual register to use for the argument.
2356    pub vreg: Reg,
2357    /// The real register into which the arg goes.
2358    pub preg: Reg,
2359}
2360
2361/// An output return value from a call instruction: the vreg that is
2362/// defined, and the preg or stack location it is constrained to (per
2363/// the ABI).
2364#[derive(Clone, Debug)]
2365pub struct CallRetPair {
2366    /// The virtual register to define from this return value.
2367    pub vreg: Writable<Reg>,
2368    /// The real register from which the return value is read.
2369    pub location: RetLocation,
2370}
2371
2372/// A location to load a return-value from after a call completes.
2373#[derive(Clone, Debug, PartialEq, Eq)]
2374pub enum RetLocation {
2375    /// A physical register.
2376    Reg(Reg, Type),
2377    /// A stack location, identified by a `StackAMode`.
2378    Stack(StackAMode, Type),
2379}
2380
2381pub type CallArgList = SmallVec<[CallArgPair; 8]>;
2382pub type CallRetList = SmallVec<[CallRetPair; 8]>;
2383
2384impl<T> CallInfo<T> {
2385    /// Emit loads for any stack-carried return values using the call
2386    /// info and allocations.
2387    pub fn emit_retval_loads<
2388        M: ABIMachineSpec,
2389        EmitFn: FnMut(M::I),
2390        IslandFn: Fn(u32) -> Option<M::I>,
2391    >(
2392        &self,
2393        stackslots_size: u32,
2394        mut emit: EmitFn,
2395        emit_island: IslandFn,
2396    ) {
2397        // Count stack-ret locations and emit an island to account for
2398        // this space usage.
2399        let mut space_needed = 0;
2400        for CallRetPair { location, .. } in &self.defs {
2401            if let RetLocation::Stack(..) = location {
2402                // Assume up to ten instructions, semi-arbitrarily:
2403                // load from stack, store to spillslot, codegen of
2404                // large offsets on RISC ISAs.
2405                space_needed += 10 * M::I::worst_case_size();
2406            }
2407        }
2408        if space_needed > 0 {
2409            if let Some(island_inst) = emit_island(space_needed) {
2410                emit(island_inst);
2411            }
2412        }
2413
2414        let temp = M::retval_temp_reg(self.callee_conv);
2415        // The temporary must be noted as clobbered.
2416        debug_assert!(
2417            M::get_regs_clobbered_by_call(self.callee_conv, self.try_call_info.is_some())
2418                .contains(PReg::from(temp.to_reg().to_real_reg().unwrap()))
2419        );
2420
2421        for CallRetPair { vreg, location } in &self.defs {
2422            match location {
2423                RetLocation::Reg(preg, ..) => {
2424                    // The temporary must not also be an actual return
2425                    // value register.
2426                    debug_assert!(*preg != temp.to_reg());
2427                }
2428                RetLocation::Stack(amode, ty) => {
2429                    if let Some(spillslot) = vreg.to_reg().to_spillslot() {
2430                        // `temp` is an integer register of machine word
2431                        // width, but `ty` may be floating-point/vector,
2432                        // which (i) may not be loadable directly into an
2433                        // int reg, and (ii) may be wider than a machine
2434                        // word. For simplicity, and because there are not
2435                        // always easy choices for volatile float/vec regs
2436                        // (see e.g. x86-64, where fastcall clobbers only
2437                        // xmm0-xmm5, but tail uses xmm0-xmm7 for
2438                        // returns), we use the integer temp register in
2439                        // steps.
2440                        let parts = (ty.bytes() + M::word_bytes() - 1) / M::word_bytes();
2441                        for part in 0..parts {
2442                            emit(M::gen_load_stack(
2443                                amode.offset_by(part * M::word_bytes()),
2444                                temp,
2445                                M::word_type(),
2446                            ));
2447                            emit(M::gen_store_stack(
2448                                StackAMode::Slot(
2449                                    i64::from(stackslots_size)
2450                                        + i64::from(M::word_bytes())
2451                                            * ((spillslot.index() as i64) + (part as i64)),
2452                                ),
2453                                temp.to_reg(),
2454                                M::word_type(),
2455                            ));
2456                        }
2457                    } else {
2458                        assert_ne!(*vreg, temp);
2459                        emit(M::gen_load_stack(*amode, *vreg, *ty));
2460                    }
2461                }
2462            }
2463        }
2464    }
2465}
2466
2467#[cfg(test)]
2468mod tests {
2469    use super::SigData;
2470
2471    #[test]
2472    fn sig_data_size() {
2473        // The size of `SigData` is performance sensitive, so make sure
2474        // we don't regress it unintentionally.
2475        assert_eq!(std::mem::size_of::<SigData>(), 24);
2476    }
2477}