X86CallLowering.cpp 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418
  1. //===- llvm/lib/Target/X86/X86CallLowering.cpp - Call lowering ------------===//
  2. //
  3. // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
  4. // See https://llvm.org/LICENSE.txt for license information.
  5. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
  6. //
  7. //===----------------------------------------------------------------------===//
  8. //
  9. /// \file
  10. /// This file implements the lowering of LLVM calls to machine code calls for
  11. /// GlobalISel.
  12. //
  13. //===----------------------------------------------------------------------===//
  14. #include "X86CallLowering.h"
  15. #include "X86CallingConv.h"
  16. #include "X86ISelLowering.h"
  17. #include "X86InstrInfo.h"
  18. #include "X86RegisterInfo.h"
  19. #include "X86Subtarget.h"
  20. #include "llvm/ADT/ArrayRef.h"
  21. #include "llvm/ADT/SmallVector.h"
  22. #include "llvm/CodeGen/Analysis.h"
  23. #include "llvm/CodeGen/CallingConvLower.h"
  24. #include "llvm/CodeGen/FunctionLoweringInfo.h"
  25. #include "llvm/CodeGen/GlobalISel/MachineIRBuilder.h"
  26. #include "llvm/CodeGen/GlobalISel/Utils.h"
  27. #include "llvm/CodeGen/LowLevelType.h"
  28. #include "llvm/CodeGen/MachineBasicBlock.h"
  29. #include "llvm/CodeGen/MachineFrameInfo.h"
  30. #include "llvm/CodeGen/MachineFunction.h"
  31. #include "llvm/CodeGen/MachineInstrBuilder.h"
  32. #include "llvm/CodeGen/MachineMemOperand.h"
  33. #include "llvm/CodeGen/MachineOperand.h"
  34. #include "llvm/CodeGen/MachineRegisterInfo.h"
  35. #include "llvm/CodeGen/TargetInstrInfo.h"
  36. #include "llvm/CodeGen/TargetSubtargetInfo.h"
  37. #include "llvm/CodeGen/ValueTypes.h"
  38. #include "llvm/IR/Attributes.h"
  39. #include "llvm/IR/DataLayout.h"
  40. #include "llvm/IR/Function.h"
  41. #include "llvm/IR/Value.h"
  42. #include "llvm/MC/MCRegisterInfo.h"
  43. #include "llvm/Support/LowLevelTypeImpl.h"
  44. #include "llvm/Support/MachineValueType.h"
  45. #include <cassert>
  46. #include <cstdint>
  47. using namespace llvm;
  48. X86CallLowering::X86CallLowering(const X86TargetLowering &TLI)
  49. : CallLowering(&TLI) {}
  50. namespace {
  51. struct X86OutgoingValueAssigner : public CallLowering::OutgoingValueAssigner {
  52. private:
  53. uint64_t StackSize = 0;
  54. unsigned NumXMMRegs = 0;
  55. public:
  56. uint64_t getStackSize() { return StackSize; }
  57. unsigned getNumXmmRegs() { return NumXMMRegs; }
  58. X86OutgoingValueAssigner(CCAssignFn *AssignFn_)
  59. : CallLowering::OutgoingValueAssigner(AssignFn_) {}
  60. bool assignArg(unsigned ValNo, EVT OrigVT, MVT ValVT, MVT LocVT,
  61. CCValAssign::LocInfo LocInfo,
  62. const CallLowering::ArgInfo &Info, ISD::ArgFlagsTy Flags,
  63. CCState &State) override {
  64. bool Res = AssignFn(ValNo, ValVT, LocVT, LocInfo, Flags, State);
  65. StackSize = State.getNextStackOffset();
  66. static const MCPhysReg XMMArgRegs[] = {X86::XMM0, X86::XMM1, X86::XMM2,
  67. X86::XMM3, X86::XMM4, X86::XMM5,
  68. X86::XMM6, X86::XMM7};
  69. if (!Info.IsFixed)
  70. NumXMMRegs = State.getFirstUnallocated(XMMArgRegs);
  71. return Res;
  72. }
  73. };
  74. struct X86OutgoingValueHandler : public CallLowering::OutgoingValueHandler {
  75. X86OutgoingValueHandler(MachineIRBuilder &MIRBuilder,
  76. MachineRegisterInfo &MRI, MachineInstrBuilder &MIB)
  77. : OutgoingValueHandler(MIRBuilder, MRI), MIB(MIB),
  78. DL(MIRBuilder.getMF().getDataLayout()),
  79. STI(MIRBuilder.getMF().getSubtarget<X86Subtarget>()) {}
  80. Register getStackAddress(uint64_t Size, int64_t Offset,
  81. MachinePointerInfo &MPO,
  82. ISD::ArgFlagsTy Flags) override {
  83. LLT p0 = LLT::pointer(0, DL.getPointerSizeInBits(0));
  84. LLT SType = LLT::scalar(DL.getPointerSizeInBits(0));
  85. auto SPReg =
  86. MIRBuilder.buildCopy(p0, STI.getRegisterInfo()->getStackRegister());
  87. auto OffsetReg = MIRBuilder.buildConstant(SType, Offset);
  88. auto AddrReg = MIRBuilder.buildPtrAdd(p0, SPReg, OffsetReg);
  89. MPO = MachinePointerInfo::getStack(MIRBuilder.getMF(), Offset);
  90. return AddrReg.getReg(0);
  91. }
  92. void assignValueToReg(Register ValVReg, Register PhysReg,
  93. CCValAssign VA) override {
  94. MIB.addUse(PhysReg, RegState::Implicit);
  95. Register ExtReg = extendRegister(ValVReg, VA);
  96. MIRBuilder.buildCopy(PhysReg, ExtReg);
  97. }
  98. void assignValueToAddress(Register ValVReg, Register Addr, LLT MemTy,
  99. MachinePointerInfo &MPO, CCValAssign &VA) override {
  100. MachineFunction &MF = MIRBuilder.getMF();
  101. Register ExtReg = extendRegister(ValVReg, VA);
  102. auto *MMO = MF.getMachineMemOperand(MPO, MachineMemOperand::MOStore, MemTy,
  103. inferAlignFromPtrInfo(MF, MPO));
  104. MIRBuilder.buildStore(ExtReg, Addr, *MMO);
  105. }
  106. protected:
  107. MachineInstrBuilder &MIB;
  108. const DataLayout &DL;
  109. const X86Subtarget &STI;
  110. };
  111. } // end anonymous namespace
  112. bool X86CallLowering::canLowerReturn(
  113. MachineFunction &MF, CallingConv::ID CallConv,
  114. SmallVectorImpl<CallLowering::BaseArgInfo> &Outs, bool IsVarArg) const {
  115. LLVMContext &Context = MF.getFunction().getContext();
  116. SmallVector<CCValAssign, 16> RVLocs;
  117. CCState CCInfo(CallConv, IsVarArg, MF, RVLocs, Context);
  118. return checkReturn(CCInfo, Outs, RetCC_X86);
  119. }
  120. bool X86CallLowering::lowerReturn(MachineIRBuilder &MIRBuilder,
  121. const Value *Val, ArrayRef<Register> VRegs,
  122. FunctionLoweringInfo &FLI) const {
  123. assert(((Val && !VRegs.empty()) || (!Val && VRegs.empty())) &&
  124. "Return value without a vreg");
  125. MachineFunction &MF = MIRBuilder.getMF();
  126. auto MIB = MIRBuilder.buildInstrNoInsert(X86::RET).addImm(0);
  127. const X86Subtarget &STI = MF.getSubtarget<X86Subtarget>();
  128. bool Is64Bit = STI.is64Bit();
  129. if (!FLI.CanLowerReturn) {
  130. insertSRetStores(MIRBuilder, Val->getType(), VRegs, FLI.DemoteRegister);
  131. MIRBuilder.buildCopy(Is64Bit ? X86::RAX : X86::EAX, FLI.DemoteRegister);
  132. } else if (!VRegs.empty()) {
  133. const Function &F = MF.getFunction();
  134. MachineRegisterInfo &MRI = MF.getRegInfo();
  135. const DataLayout &DL = MF.getDataLayout();
  136. ArgInfo OrigRetInfo(VRegs, Val->getType(), 0);
  137. setArgFlags(OrigRetInfo, AttributeList::ReturnIndex, DL, F);
  138. SmallVector<ArgInfo, 4> SplitRetInfos;
  139. splitToValueTypes(OrigRetInfo, SplitRetInfos, DL, F.getCallingConv());
  140. X86OutgoingValueAssigner Assigner(RetCC_X86);
  141. X86OutgoingValueHandler Handler(MIRBuilder, MRI, MIB);
  142. if (!determineAndHandleAssignments(Handler, Assigner, SplitRetInfos,
  143. MIRBuilder, F.getCallingConv(),
  144. F.isVarArg()))
  145. return false;
  146. }
  147. MIRBuilder.insertInstr(MIB);
  148. return true;
  149. }
  150. namespace {
  151. struct X86IncomingValueHandler : public CallLowering::IncomingValueHandler {
  152. X86IncomingValueHandler(MachineIRBuilder &MIRBuilder,
  153. MachineRegisterInfo &MRI)
  154. : IncomingValueHandler(MIRBuilder, MRI),
  155. DL(MIRBuilder.getMF().getDataLayout()) {}
  156. Register getStackAddress(uint64_t Size, int64_t Offset,
  157. MachinePointerInfo &MPO,
  158. ISD::ArgFlagsTy Flags) override {
  159. auto &MFI = MIRBuilder.getMF().getFrameInfo();
  160. // Byval is assumed to be writable memory, but other stack passed arguments
  161. // are not.
  162. const bool IsImmutable = !Flags.isByVal();
  163. int FI = MFI.CreateFixedObject(Size, Offset, IsImmutable);
  164. MPO = MachinePointerInfo::getFixedStack(MIRBuilder.getMF(), FI);
  165. return MIRBuilder
  166. .buildFrameIndex(LLT::pointer(0, DL.getPointerSizeInBits(0)), FI)
  167. .getReg(0);
  168. }
  169. void assignValueToAddress(Register ValVReg, Register Addr, LLT MemTy,
  170. MachinePointerInfo &MPO, CCValAssign &VA) override {
  171. MachineFunction &MF = MIRBuilder.getMF();
  172. auto *MMO = MF.getMachineMemOperand(
  173. MPO, MachineMemOperand::MOLoad | MachineMemOperand::MOInvariant, MemTy,
  174. inferAlignFromPtrInfo(MF, MPO));
  175. MIRBuilder.buildLoad(ValVReg, Addr, *MMO);
  176. }
  177. void assignValueToReg(Register ValVReg, Register PhysReg,
  178. CCValAssign VA) override {
  179. markPhysRegUsed(PhysReg);
  180. IncomingValueHandler::assignValueToReg(ValVReg, PhysReg, VA);
  181. }
  182. /// How the physical register gets marked varies between formal
  183. /// parameters (it's a basic-block live-in), and a call instruction
  184. /// (it's an implicit-def of the BL).
  185. virtual void markPhysRegUsed(unsigned PhysReg) = 0;
  186. protected:
  187. const DataLayout &DL;
  188. };
  189. struct FormalArgHandler : public X86IncomingValueHandler {
  190. FormalArgHandler(MachineIRBuilder &MIRBuilder, MachineRegisterInfo &MRI)
  191. : X86IncomingValueHandler(MIRBuilder, MRI) {}
  192. void markPhysRegUsed(unsigned PhysReg) override {
  193. MIRBuilder.getMRI()->addLiveIn(PhysReg);
  194. MIRBuilder.getMBB().addLiveIn(PhysReg);
  195. }
  196. };
  197. struct CallReturnHandler : public X86IncomingValueHandler {
  198. CallReturnHandler(MachineIRBuilder &MIRBuilder, MachineRegisterInfo &MRI,
  199. MachineInstrBuilder &MIB)
  200. : X86IncomingValueHandler(MIRBuilder, MRI), MIB(MIB) {}
  201. void markPhysRegUsed(unsigned PhysReg) override {
  202. MIB.addDef(PhysReg, RegState::Implicit);
  203. }
  204. protected:
  205. MachineInstrBuilder &MIB;
  206. };
  207. } // end anonymous namespace
  208. bool X86CallLowering::lowerFormalArguments(MachineIRBuilder &MIRBuilder,
  209. const Function &F,
  210. ArrayRef<ArrayRef<Register>> VRegs,
  211. FunctionLoweringInfo &FLI) const {
  212. MachineFunction &MF = MIRBuilder.getMF();
  213. MachineRegisterInfo &MRI = MF.getRegInfo();
  214. auto DL = MF.getDataLayout();
  215. SmallVector<ArgInfo, 8> SplitArgs;
  216. if (!FLI.CanLowerReturn)
  217. insertSRetIncomingArgument(F, SplitArgs, FLI.DemoteRegister, MRI, DL);
  218. // TODO: handle variadic function
  219. if (F.isVarArg())
  220. return false;
  221. unsigned Idx = 0;
  222. for (const auto &Arg : F.args()) {
  223. // TODO: handle not simple cases.
  224. if (Arg.hasAttribute(Attribute::ByVal) ||
  225. Arg.hasAttribute(Attribute::InReg) ||
  226. Arg.hasAttribute(Attribute::StructRet) ||
  227. Arg.hasAttribute(Attribute::SwiftSelf) ||
  228. Arg.hasAttribute(Attribute::SwiftError) ||
  229. Arg.hasAttribute(Attribute::Nest) || VRegs[Idx].size() > 1)
  230. return false;
  231. ArgInfo OrigArg(VRegs[Idx], Arg.getType(), Idx);
  232. setArgFlags(OrigArg, Idx + AttributeList::FirstArgIndex, DL, F);
  233. splitToValueTypes(OrigArg, SplitArgs, DL, F.getCallingConv());
  234. Idx++;
  235. }
  236. if (SplitArgs.empty())
  237. return true;
  238. MachineBasicBlock &MBB = MIRBuilder.getMBB();
  239. if (!MBB.empty())
  240. MIRBuilder.setInstr(*MBB.begin());
  241. X86OutgoingValueAssigner Assigner(CC_X86);
  242. FormalArgHandler Handler(MIRBuilder, MRI);
  243. if (!determineAndHandleAssignments(Handler, Assigner, SplitArgs, MIRBuilder,
  244. F.getCallingConv(), F.isVarArg()))
  245. return false;
  246. // Move back to the end of the basic block.
  247. MIRBuilder.setMBB(MBB);
  248. return true;
  249. }
  250. bool X86CallLowering::lowerCall(MachineIRBuilder &MIRBuilder,
  251. CallLoweringInfo &Info) const {
  252. MachineFunction &MF = MIRBuilder.getMF();
  253. const Function &F = MF.getFunction();
  254. MachineRegisterInfo &MRI = MF.getRegInfo();
  255. const DataLayout &DL = F.getParent()->getDataLayout();
  256. const X86Subtarget &STI = MF.getSubtarget<X86Subtarget>();
  257. const TargetInstrInfo &TII = *STI.getInstrInfo();
  258. const X86RegisterInfo *TRI = STI.getRegisterInfo();
  259. // Handle only Linux C, X86_64_SysV calling conventions for now.
  260. if (!STI.isTargetLinux() || !(Info.CallConv == CallingConv::C ||
  261. Info.CallConv == CallingConv::X86_64_SysV))
  262. return false;
  263. unsigned AdjStackDown = TII.getCallFrameSetupOpcode();
  264. auto CallSeqStart = MIRBuilder.buildInstr(AdjStackDown);
  265. // Create a temporarily-floating call instruction so we can add the implicit
  266. // uses of arg registers.
  267. bool Is64Bit = STI.is64Bit();
  268. unsigned CallOpc = Info.Callee.isReg()
  269. ? (Is64Bit ? X86::CALL64r : X86::CALL32r)
  270. : (Is64Bit ? X86::CALL64pcrel32 : X86::CALLpcrel32);
  271. auto MIB = MIRBuilder.buildInstrNoInsert(CallOpc)
  272. .add(Info.Callee)
  273. .addRegMask(TRI->getCallPreservedMask(MF, Info.CallConv));
  274. SmallVector<ArgInfo, 8> SplitArgs;
  275. for (const auto &OrigArg : Info.OrigArgs) {
  276. // TODO: handle not simple cases.
  277. if (OrigArg.Flags[0].isByVal())
  278. return false;
  279. if (OrigArg.Regs.size() > 1)
  280. return false;
  281. splitToValueTypes(OrigArg, SplitArgs, DL, Info.CallConv);
  282. }
  283. // Do the actual argument marshalling.
  284. X86OutgoingValueAssigner Assigner(CC_X86);
  285. X86OutgoingValueHandler Handler(MIRBuilder, MRI, MIB);
  286. if (!determineAndHandleAssignments(Handler, Assigner, SplitArgs, MIRBuilder,
  287. Info.CallConv, Info.IsVarArg))
  288. return false;
  289. bool IsFixed = Info.OrigArgs.empty() ? true : Info.OrigArgs.back().IsFixed;
  290. if (STI.is64Bit() && !IsFixed && !STI.isCallingConvWin64(Info.CallConv)) {
  291. // From AMD64 ABI document:
  292. // For calls that may call functions that use varargs or stdargs
  293. // (prototype-less calls or calls to functions containing ellipsis (...) in
  294. // the declaration) %al is used as hidden argument to specify the number
  295. // of SSE registers used. The contents of %al do not need to match exactly
  296. // the number of registers, but must be an ubound on the number of SSE
  297. // registers used and is in the range 0 - 8 inclusive.
  298. MIRBuilder.buildInstr(X86::MOV8ri)
  299. .addDef(X86::AL)
  300. .addImm(Assigner.getNumXmmRegs());
  301. MIB.addUse(X86::AL, RegState::Implicit);
  302. }
  303. // Now we can add the actual call instruction to the correct basic block.
  304. MIRBuilder.insertInstr(MIB);
  305. // If Callee is a reg, since it is used by a target specific
  306. // instruction, it must have a register class matching the
  307. // constraint of that instruction.
  308. if (Info.Callee.isReg())
  309. MIB->getOperand(0).setReg(constrainOperandRegClass(
  310. MF, *TRI, MRI, *MF.getSubtarget().getInstrInfo(),
  311. *MF.getSubtarget().getRegBankInfo(), *MIB, MIB->getDesc(), Info.Callee,
  312. 0));
  313. // Finally we can copy the returned value back into its virtual-register. In
  314. // symmetry with the arguments, the physical register must be an
  315. // implicit-define of the call instruction.
  316. if (Info.CanLowerReturn && !Info.OrigRet.Ty->isVoidTy()) {
  317. if (Info.OrigRet.Regs.size() > 1)
  318. return false;
  319. SplitArgs.clear();
  320. SmallVector<Register, 8> NewRegs;
  321. splitToValueTypes(Info.OrigRet, SplitArgs, DL, Info.CallConv);
  322. X86OutgoingValueAssigner Assigner(RetCC_X86);
  323. CallReturnHandler Handler(MIRBuilder, MRI, MIB);
  324. if (!determineAndHandleAssignments(Handler, Assigner, SplitArgs, MIRBuilder,
  325. Info.CallConv, Info.IsVarArg))
  326. return false;
  327. if (!NewRegs.empty())
  328. MIRBuilder.buildMergeLikeInstr(Info.OrigRet.Regs[0], NewRegs);
  329. }
  330. CallSeqStart.addImm(Assigner.getStackSize())
  331. .addImm(0 /* see getFrameTotalSize */)
  332. .addImm(0 /* see getFrameAdjustment */);
  333. unsigned AdjStackUp = TII.getCallFrameDestroyOpcode();
  334. MIRBuilder.buildInstr(AdjStackUp)
  335. .addImm(Assigner.getStackSize())
  336. .addImm(0 /* NumBytesForCalleeToPop */);
  337. if (!Info.CanLowerReturn)
  338. insertSRetLoads(MIRBuilder, Info.OrigRet.Ty, Info.OrigRet.Regs,
  339. Info.DemoteRegister, Info.DemoteStackIndex);
  340. return true;
  341. }