123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207 |
- //=== AArch64CallingConvention.cpp - AArch64 CC impl ------------*- C++ -*-===//
- //
- // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
- // See https://llvm.org/LICENSE.txt for license information.
- // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
- //
- //===----------------------------------------------------------------------===//
- //
- // This file contains the table-generated and custom routines for the AArch64
- // Calling Convention.
- //
- //===----------------------------------------------------------------------===//
- #include "AArch64CallingConvention.h"
- #include "AArch64.h"
- #include "AArch64InstrInfo.h"
- #include "AArch64Subtarget.h"
- #include "llvm/CodeGen/CallingConvLower.h"
- #include "llvm/CodeGen/TargetInstrInfo.h"
- #include "llvm/IR/CallingConv.h"
- using namespace llvm;
- static const MCPhysReg XRegList[] = {AArch64::X0, AArch64::X1, AArch64::X2,
- AArch64::X3, AArch64::X4, AArch64::X5,
- AArch64::X6, AArch64::X7};
- static const MCPhysReg HRegList[] = {AArch64::H0, AArch64::H1, AArch64::H2,
- AArch64::H3, AArch64::H4, AArch64::H5,
- AArch64::H6, AArch64::H7};
- static const MCPhysReg SRegList[] = {AArch64::S0, AArch64::S1, AArch64::S2,
- AArch64::S3, AArch64::S4, AArch64::S5,
- AArch64::S6, AArch64::S7};
- static const MCPhysReg DRegList[] = {AArch64::D0, AArch64::D1, AArch64::D2,
- AArch64::D3, AArch64::D4, AArch64::D5,
- AArch64::D6, AArch64::D7};
- static const MCPhysReg QRegList[] = {AArch64::Q0, AArch64::Q1, AArch64::Q2,
- AArch64::Q3, AArch64::Q4, AArch64::Q5,
- AArch64::Q6, AArch64::Q7};
- static const MCPhysReg ZRegList[] = {AArch64::Z0, AArch64::Z1, AArch64::Z2,
- AArch64::Z3, AArch64::Z4, AArch64::Z5,
- AArch64::Z6, AArch64::Z7};
- static bool finishStackBlock(SmallVectorImpl<CCValAssign> &PendingMembers,
- MVT LocVT, ISD::ArgFlagsTy &ArgFlags,
- CCState &State, Align SlotAlign) {
- if (LocVT.isScalableVector()) {
- const AArch64Subtarget &Subtarget = static_cast<const AArch64Subtarget &>(
- State.getMachineFunction().getSubtarget());
- const AArch64TargetLowering *TLI = Subtarget.getTargetLowering();
- // We are about to reinvoke the CCAssignFn auto-generated handler. If we
- // don't unset these flags we will get stuck in an infinite loop forever
- // invoking the custom handler.
- ArgFlags.setInConsecutiveRegs(false);
- ArgFlags.setInConsecutiveRegsLast(false);
- // The calling convention for passing SVE tuples states that in the event
- // we cannot allocate enough registers for the tuple we should still leave
- // any remaining registers unallocated. However, when we call the
- // CCAssignFn again we want it to behave as if all remaining registers are
- // allocated. This will force the code to pass the tuple indirectly in
- // accordance with the PCS.
- bool RegsAllocated[8];
- for (int I = 0; I < 8; I++) {
- RegsAllocated[I] = State.isAllocated(ZRegList[I]);
- State.AllocateReg(ZRegList[I]);
- }
- auto &It = PendingMembers[0];
- CCAssignFn *AssignFn =
- TLI->CCAssignFnForCall(State.getCallingConv(), /*IsVarArg=*/false);
- if (AssignFn(It.getValNo(), It.getValVT(), It.getValVT(), CCValAssign::Full,
- ArgFlags, State))
- llvm_unreachable("Call operand has unhandled type");
- // Return the flags to how they were before.
- ArgFlags.setInConsecutiveRegs(true);
- ArgFlags.setInConsecutiveRegsLast(true);
- // Return the register state back to how it was before, leaving any
- // unallocated registers available for other smaller types.
- for (int I = 0; I < 8; I++)
- if (!RegsAllocated[I])
- State.DeallocateReg(ZRegList[I]);
- // All pending members have now been allocated
- PendingMembers.clear();
- return true;
- }
- unsigned Size = LocVT.getSizeInBits() / 8;
- for (auto &It : PendingMembers) {
- It.convertToMem(State.AllocateStack(Size, SlotAlign));
- State.addLoc(It);
- SlotAlign = Align(1);
- }
- // All pending members have now been allocated
- PendingMembers.clear();
- return true;
- }
- /// The Darwin variadic PCS places anonymous arguments in 8-byte stack slots. An
- /// [N x Ty] type must still be contiguous in memory though.
- static bool CC_AArch64_Custom_Stack_Block(
- unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo,
- ISD::ArgFlagsTy &ArgFlags, CCState &State) {
- SmallVectorImpl<CCValAssign> &PendingMembers = State.getPendingLocs();
- // Add the argument to the list to be allocated once we know the size of the
- // block.
- PendingMembers.push_back(
- CCValAssign::getPending(ValNo, ValVT, LocVT, LocInfo));
- if (!ArgFlags.isInConsecutiveRegsLast())
- return true;
- return finishStackBlock(PendingMembers, LocVT, ArgFlags, State, Align(8));
- }
- /// Given an [N x Ty] block, it should be passed in a consecutive sequence of
- /// registers. If no such sequence is available, mark the rest of the registers
- /// of that type as used and place the argument on the stack.
- static bool CC_AArch64_Custom_Block(unsigned &ValNo, MVT &ValVT, MVT &LocVT,
- CCValAssign::LocInfo &LocInfo,
- ISD::ArgFlagsTy &ArgFlags, CCState &State) {
- const AArch64Subtarget &Subtarget = static_cast<const AArch64Subtarget &>(
- State.getMachineFunction().getSubtarget());
- bool IsDarwinILP32 = Subtarget.isTargetILP32() && Subtarget.isTargetMachO();
- // Try to allocate a contiguous block of registers, each of the correct
- // size to hold one member.
- ArrayRef<MCPhysReg> RegList;
- if (LocVT.SimpleTy == MVT::i64 || (IsDarwinILP32 && LocVT.SimpleTy == MVT::i32))
- RegList = XRegList;
- else if (LocVT.SimpleTy == MVT::f16)
- RegList = HRegList;
- else if (LocVT.SimpleTy == MVT::f32 || LocVT.is32BitVector())
- RegList = SRegList;
- else if (LocVT.SimpleTy == MVT::f64 || LocVT.is64BitVector())
- RegList = DRegList;
- else if (LocVT.SimpleTy == MVT::f128 || LocVT.is128BitVector())
- RegList = QRegList;
- else if (LocVT.isScalableVector())
- RegList = ZRegList;
- else {
- // Not an array we want to split up after all.
- return false;
- }
- SmallVectorImpl<CCValAssign> &PendingMembers = State.getPendingLocs();
- // Add the argument to the list to be allocated once we know the size of the
- // block.
- PendingMembers.push_back(
- CCValAssign::getPending(ValNo, ValVT, LocVT, LocInfo));
- if (!ArgFlags.isInConsecutiveRegsLast())
- return true;
- // [N x i32] arguments get packed into x-registers on Darwin's arm64_32
- // because that's how the armv7k Clang front-end emits small structs.
- unsigned EltsPerReg = (IsDarwinILP32 && LocVT.SimpleTy == MVT::i32) ? 2 : 1;
- unsigned RegResult = State.AllocateRegBlock(
- RegList, alignTo(PendingMembers.size(), EltsPerReg) / EltsPerReg);
- if (RegResult && EltsPerReg == 1) {
- for (auto &It : PendingMembers) {
- It.convertToReg(RegResult);
- State.addLoc(It);
- ++RegResult;
- }
- PendingMembers.clear();
- return true;
- } else if (RegResult) {
- assert(EltsPerReg == 2 && "unexpected ABI");
- bool UseHigh = false;
- CCValAssign::LocInfo Info;
- for (auto &It : PendingMembers) {
- Info = UseHigh ? CCValAssign::AExtUpper : CCValAssign::ZExt;
- State.addLoc(CCValAssign::getReg(It.getValNo(), MVT::i32, RegResult,
- MVT::i64, Info));
- UseHigh = !UseHigh;
- if (!UseHigh)
- ++RegResult;
- }
- PendingMembers.clear();
- return true;
- }
- if (!LocVT.isScalableVector()) {
- // Mark all regs in the class as unavailable
- for (auto Reg : RegList)
- State.AllocateReg(Reg);
- }
- const Align StackAlign =
- State.getMachineFunction().getDataLayout().getStackAlignment();
- const Align MemAlign = ArgFlags.getNonZeroMemAlign();
- Align SlotAlign = std::min(MemAlign, StackAlign);
- if (!Subtarget.isTargetDarwin())
- SlotAlign = std::max(SlotAlign, Align(8));
- return finishStackBlock(PendingMembers, LocVT, ArgFlags, State, SlotAlign);
- }
- // TableGen provides definitions of the calling convention analysis entry
- // points.
- #include "AArch64GenCallingConv.inc"
|