AArch64PostLegalizerCombiner.cpp 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451
  1. //=== AArch64PostLegalizerCombiner.cpp --------------------------*- C++ -*-===//
  2. //
  3. // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
  4. // See https://llvm.org/LICENSE.txt for license information.
  5. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
  6. //
  7. //===----------------------------------------------------------------------===//
  8. ///
  9. /// \file
  10. /// Post-legalization combines on generic MachineInstrs.
  11. ///
  12. /// The combines here must preserve instruction legality.
  13. ///
  14. /// Lowering combines (e.g. pseudo matching) should be handled by
  15. /// AArch64PostLegalizerLowering.
  16. ///
  17. /// Combines which don't rely on instruction legality should go in the
  18. /// AArch64PreLegalizerCombiner.
  19. ///
  20. //===----------------------------------------------------------------------===//
  21. #include "AArch64TargetMachine.h"
  22. #include "llvm/CodeGen/GlobalISel/CSEInfo.h"
  23. #include "llvm/CodeGen/GlobalISel/Combiner.h"
  24. #include "llvm/CodeGen/GlobalISel/CombinerHelper.h"
  25. #include "llvm/CodeGen/GlobalISel/CombinerInfo.h"
  26. #include "llvm/CodeGen/GlobalISel/GISelChangeObserver.h"
  27. #include "llvm/CodeGen/GlobalISel/GISelKnownBits.h"
  28. #include "llvm/CodeGen/GlobalISel/GenericMachineInstrs.h"
  29. #include "llvm/CodeGen/GlobalISel/MIPatternMatch.h"
  30. #include "llvm/CodeGen/GlobalISel/MachineIRBuilder.h"
  31. #include "llvm/CodeGen/GlobalISel/Utils.h"
  32. #include "llvm/CodeGen/MachineDominators.h"
  33. #include "llvm/CodeGen/MachineFunctionPass.h"
  34. #include "llvm/CodeGen/MachineRegisterInfo.h"
  35. #include "llvm/CodeGen/TargetOpcodes.h"
  36. #include "llvm/CodeGen/TargetPassConfig.h"
  37. #include "llvm/Support/Debug.h"
  38. #define DEBUG_TYPE "aarch64-postlegalizer-combiner"
  39. using namespace llvm;
  40. using namespace MIPatternMatch;
  41. /// This combine tries do what performExtractVectorEltCombine does in SDAG.
  42. /// Rewrite for pairwise fadd pattern
  43. /// (s32 (g_extract_vector_elt
  44. /// (g_fadd (vXs32 Other)
  45. /// (g_vector_shuffle (vXs32 Other) undef <1,X,...> )) 0))
  46. /// ->
  47. /// (s32 (g_fadd (g_extract_vector_elt (vXs32 Other) 0)
  48. /// (g_extract_vector_elt (vXs32 Other) 1))
  49. bool matchExtractVecEltPairwiseAdd(
  50. MachineInstr &MI, MachineRegisterInfo &MRI,
  51. std::tuple<unsigned, LLT, Register> &MatchInfo) {
  52. Register Src1 = MI.getOperand(1).getReg();
  53. Register Src2 = MI.getOperand(2).getReg();
  54. LLT DstTy = MRI.getType(MI.getOperand(0).getReg());
  55. auto Cst = getIConstantVRegValWithLookThrough(Src2, MRI);
  56. if (!Cst || Cst->Value != 0)
  57. return false;
  58. // SDAG also checks for FullFP16, but this looks to be beneficial anyway.
  59. // Now check for an fadd operation. TODO: expand this for integer add?
  60. auto *FAddMI = getOpcodeDef(TargetOpcode::G_FADD, Src1, MRI);
  61. if (!FAddMI)
  62. return false;
  63. // If we add support for integer add, must restrict these types to just s64.
  64. unsigned DstSize = DstTy.getSizeInBits();
  65. if (DstSize != 16 && DstSize != 32 && DstSize != 64)
  66. return false;
  67. Register Src1Op1 = FAddMI->getOperand(1).getReg();
  68. Register Src1Op2 = FAddMI->getOperand(2).getReg();
  69. MachineInstr *Shuffle =
  70. getOpcodeDef(TargetOpcode::G_SHUFFLE_VECTOR, Src1Op2, MRI);
  71. MachineInstr *Other = MRI.getVRegDef(Src1Op1);
  72. if (!Shuffle) {
  73. Shuffle = getOpcodeDef(TargetOpcode::G_SHUFFLE_VECTOR, Src1Op1, MRI);
  74. Other = MRI.getVRegDef(Src1Op2);
  75. }
  76. // We're looking for a shuffle that moves the second element to index 0.
  77. if (Shuffle && Shuffle->getOperand(3).getShuffleMask()[0] == 1 &&
  78. Other == MRI.getVRegDef(Shuffle->getOperand(1).getReg())) {
  79. std::get<0>(MatchInfo) = TargetOpcode::G_FADD;
  80. std::get<1>(MatchInfo) = DstTy;
  81. std::get<2>(MatchInfo) = Other->getOperand(0).getReg();
  82. return true;
  83. }
  84. return false;
  85. }
  86. bool applyExtractVecEltPairwiseAdd(
  87. MachineInstr &MI, MachineRegisterInfo &MRI, MachineIRBuilder &B,
  88. std::tuple<unsigned, LLT, Register> &MatchInfo) {
  89. unsigned Opc = std::get<0>(MatchInfo);
  90. assert(Opc == TargetOpcode::G_FADD && "Unexpected opcode!");
  91. // We want to generate two extracts of elements 0 and 1, and add them.
  92. LLT Ty = std::get<1>(MatchInfo);
  93. Register Src = std::get<2>(MatchInfo);
  94. LLT s64 = LLT::scalar(64);
  95. B.setInstrAndDebugLoc(MI);
  96. auto Elt0 = B.buildExtractVectorElement(Ty, Src, B.buildConstant(s64, 0));
  97. auto Elt1 = B.buildExtractVectorElement(Ty, Src, B.buildConstant(s64, 1));
  98. B.buildInstr(Opc, {MI.getOperand(0).getReg()}, {Elt0, Elt1});
  99. MI.eraseFromParent();
  100. return true;
  101. }
  102. static bool isSignExtended(Register R, MachineRegisterInfo &MRI) {
  103. // TODO: check if extended build vector as well.
  104. unsigned Opc = MRI.getVRegDef(R)->getOpcode();
  105. return Opc == TargetOpcode::G_SEXT || Opc == TargetOpcode::G_SEXT_INREG;
  106. }
  107. static bool isZeroExtended(Register R, MachineRegisterInfo &MRI) {
  108. // TODO: check if extended build vector as well.
  109. return MRI.getVRegDef(R)->getOpcode() == TargetOpcode::G_ZEXT;
  110. }
  111. bool matchAArch64MulConstCombine(
  112. MachineInstr &MI, MachineRegisterInfo &MRI,
  113. std::function<void(MachineIRBuilder &B, Register DstReg)> &ApplyFn) {
  114. assert(MI.getOpcode() == TargetOpcode::G_MUL);
  115. Register LHS = MI.getOperand(1).getReg();
  116. Register RHS = MI.getOperand(2).getReg();
  117. Register Dst = MI.getOperand(0).getReg();
  118. const LLT Ty = MRI.getType(LHS);
  119. // The below optimizations require a constant RHS.
  120. auto Const = getIConstantVRegValWithLookThrough(RHS, MRI);
  121. if (!Const)
  122. return false;
  123. APInt ConstValue = Const->Value.sext(Ty.getSizeInBits());
  124. // The following code is ported from AArch64ISelLowering.
  125. // Multiplication of a power of two plus/minus one can be done more
  126. // cheaply as as shift+add/sub. For now, this is true unilaterally. If
  127. // future CPUs have a cheaper MADD instruction, this may need to be
  128. // gated on a subtarget feature. For Cyclone, 32-bit MADD is 4 cycles and
  129. // 64-bit is 5 cycles, so this is always a win.
  130. // More aggressively, some multiplications N0 * C can be lowered to
  131. // shift+add+shift if the constant C = A * B where A = 2^N + 1 and B = 2^M,
  132. // e.g. 6=3*2=(2+1)*2.
  133. // TODO: consider lowering more cases, e.g. C = 14, -6, -14 or even 45
  134. // which equals to (1+2)*16-(1+2).
  135. // TrailingZeroes is used to test if the mul can be lowered to
  136. // shift+add+shift.
  137. unsigned TrailingZeroes = ConstValue.countTrailingZeros();
  138. if (TrailingZeroes) {
  139. // Conservatively do not lower to shift+add+shift if the mul might be
  140. // folded into smul or umul.
  141. if (MRI.hasOneNonDBGUse(LHS) &&
  142. (isSignExtended(LHS, MRI) || isZeroExtended(LHS, MRI)))
  143. return false;
  144. // Conservatively do not lower to shift+add+shift if the mul might be
  145. // folded into madd or msub.
  146. if (MRI.hasOneNonDBGUse(Dst)) {
  147. MachineInstr &UseMI = *MRI.use_instr_begin(Dst);
  148. unsigned UseOpc = UseMI.getOpcode();
  149. if (UseOpc == TargetOpcode::G_ADD || UseOpc == TargetOpcode::G_PTR_ADD ||
  150. UseOpc == TargetOpcode::G_SUB)
  151. return false;
  152. }
  153. }
  154. // Use ShiftedConstValue instead of ConstValue to support both shift+add/sub
  155. // and shift+add+shift.
  156. APInt ShiftedConstValue = ConstValue.ashr(TrailingZeroes);
  157. unsigned ShiftAmt, AddSubOpc;
  158. // Is the shifted value the LHS operand of the add/sub?
  159. bool ShiftValUseIsLHS = true;
  160. // Do we need to negate the result?
  161. bool NegateResult = false;
  162. if (ConstValue.isNonNegative()) {
  163. // (mul x, 2^N + 1) => (add (shl x, N), x)
  164. // (mul x, 2^N - 1) => (sub (shl x, N), x)
  165. // (mul x, (2^N + 1) * 2^M) => (shl (add (shl x, N), x), M)
  166. APInt SCVMinus1 = ShiftedConstValue - 1;
  167. APInt CVPlus1 = ConstValue + 1;
  168. if (SCVMinus1.isPowerOf2()) {
  169. ShiftAmt = SCVMinus1.logBase2();
  170. AddSubOpc = TargetOpcode::G_ADD;
  171. } else if (CVPlus1.isPowerOf2()) {
  172. ShiftAmt = CVPlus1.logBase2();
  173. AddSubOpc = TargetOpcode::G_SUB;
  174. } else
  175. return false;
  176. } else {
  177. // (mul x, -(2^N - 1)) => (sub x, (shl x, N))
  178. // (mul x, -(2^N + 1)) => - (add (shl x, N), x)
  179. APInt CVNegPlus1 = -ConstValue + 1;
  180. APInt CVNegMinus1 = -ConstValue - 1;
  181. if (CVNegPlus1.isPowerOf2()) {
  182. ShiftAmt = CVNegPlus1.logBase2();
  183. AddSubOpc = TargetOpcode::G_SUB;
  184. ShiftValUseIsLHS = false;
  185. } else if (CVNegMinus1.isPowerOf2()) {
  186. ShiftAmt = CVNegMinus1.logBase2();
  187. AddSubOpc = TargetOpcode::G_ADD;
  188. NegateResult = true;
  189. } else
  190. return false;
  191. }
  192. if (NegateResult && TrailingZeroes)
  193. return false;
  194. ApplyFn = [=](MachineIRBuilder &B, Register DstReg) {
  195. auto Shift = B.buildConstant(LLT::scalar(64), ShiftAmt);
  196. auto ShiftedVal = B.buildShl(Ty, LHS, Shift);
  197. Register AddSubLHS = ShiftValUseIsLHS ? ShiftedVal.getReg(0) : LHS;
  198. Register AddSubRHS = ShiftValUseIsLHS ? LHS : ShiftedVal.getReg(0);
  199. auto Res = B.buildInstr(AddSubOpc, {Ty}, {AddSubLHS, AddSubRHS});
  200. assert(!(NegateResult && TrailingZeroes) &&
  201. "NegateResult and TrailingZeroes cannot both be true for now.");
  202. // Negate the result.
  203. if (NegateResult) {
  204. B.buildSub(DstReg, B.buildConstant(Ty, 0), Res);
  205. return;
  206. }
  207. // Shift the result.
  208. if (TrailingZeroes) {
  209. B.buildShl(DstReg, Res, B.buildConstant(LLT::scalar(64), TrailingZeroes));
  210. return;
  211. }
  212. B.buildCopy(DstReg, Res.getReg(0));
  213. };
  214. return true;
  215. }
  216. bool applyAArch64MulConstCombine(
  217. MachineInstr &MI, MachineRegisterInfo &MRI, MachineIRBuilder &B,
  218. std::function<void(MachineIRBuilder &B, Register DstReg)> &ApplyFn) {
  219. B.setInstrAndDebugLoc(MI);
  220. ApplyFn(B, MI.getOperand(0).getReg());
  221. MI.eraseFromParent();
  222. return true;
  223. }
  224. /// Try to fold a G_MERGE_VALUES of 2 s32 sources, where the second source
  225. /// is a zero, into a G_ZEXT of the first.
  226. bool matchFoldMergeToZext(MachineInstr &MI, MachineRegisterInfo &MRI) {
  227. auto &Merge = cast<GMerge>(MI);
  228. LLT SrcTy = MRI.getType(Merge.getSourceReg(0));
  229. if (SrcTy != LLT::scalar(32) || Merge.getNumSources() != 2)
  230. return false;
  231. return mi_match(Merge.getSourceReg(1), MRI, m_SpecificICst(0));
  232. }
  233. void applyFoldMergeToZext(MachineInstr &MI, MachineRegisterInfo &MRI,
  234. MachineIRBuilder &B, GISelChangeObserver &Observer) {
  235. // Mutate %d(s64) = G_MERGE_VALUES %a(s32), 0(s32)
  236. // ->
  237. // %d(s64) = G_ZEXT %a(s32)
  238. Observer.changingInstr(MI);
  239. MI.setDesc(B.getTII().get(TargetOpcode::G_ZEXT));
  240. MI.removeOperand(2);
  241. Observer.changedInstr(MI);
  242. }
  243. /// \returns True if a G_ANYEXT instruction \p MI should be mutated to a G_ZEXT
  244. /// instruction.
  245. static bool matchMutateAnyExtToZExt(MachineInstr &MI, MachineRegisterInfo &MRI) {
  246. // If this is coming from a scalar compare then we can use a G_ZEXT instead of
  247. // a G_ANYEXT:
  248. //
  249. // %cmp:_(s32) = G_[I|F]CMP ... <-- produces 0/1.
  250. // %ext:_(s64) = G_ANYEXT %cmp(s32)
  251. //
  252. // By doing this, we can leverage more KnownBits combines.
  253. assert(MI.getOpcode() == TargetOpcode::G_ANYEXT);
  254. Register Dst = MI.getOperand(0).getReg();
  255. Register Src = MI.getOperand(1).getReg();
  256. return MRI.getType(Dst).isScalar() &&
  257. mi_match(Src, MRI,
  258. m_any_of(m_GICmp(m_Pred(), m_Reg(), m_Reg()),
  259. m_GFCmp(m_Pred(), m_Reg(), m_Reg())));
  260. }
  261. static void applyMutateAnyExtToZExt(MachineInstr &MI, MachineRegisterInfo &MRI,
  262. MachineIRBuilder &B,
  263. GISelChangeObserver &Observer) {
  264. Observer.changingInstr(MI);
  265. MI.setDesc(B.getTII().get(TargetOpcode::G_ZEXT));
  266. Observer.changedInstr(MI);
  267. }
  268. /// Match a 128b store of zero and split it into two 64 bit stores, for
  269. /// size/performance reasons.
  270. static bool matchSplitStoreZero128(MachineInstr &MI, MachineRegisterInfo &MRI) {
  271. GStore &Store = cast<GStore>(MI);
  272. if (!Store.isSimple())
  273. return false;
  274. LLT ValTy = MRI.getType(Store.getValueReg());
  275. if (!ValTy.isVector() || ValTy.getSizeInBits() != 128)
  276. return false;
  277. if (ValTy.getSizeInBits() != Store.getMemSizeInBits())
  278. return false; // Don't split truncating stores.
  279. if (!MRI.hasOneNonDBGUse(Store.getValueReg()))
  280. return false;
  281. auto MaybeCst = isConstantOrConstantSplatVector(
  282. *MRI.getVRegDef(Store.getValueReg()), MRI);
  283. return MaybeCst && MaybeCst->isZero();
  284. }
  285. static void applySplitStoreZero128(MachineInstr &MI, MachineRegisterInfo &MRI,
  286. MachineIRBuilder &B,
  287. GISelChangeObserver &Observer) {
  288. B.setInstrAndDebugLoc(MI);
  289. GStore &Store = cast<GStore>(MI);
  290. assert(MRI.getType(Store.getValueReg()).isVector() &&
  291. "Expected a vector store value");
  292. LLT NewTy = LLT::scalar(64);
  293. Register PtrReg = Store.getPointerReg();
  294. auto Zero = B.buildConstant(NewTy, 0);
  295. auto HighPtr = B.buildPtrAdd(MRI.getType(PtrReg), PtrReg,
  296. B.buildConstant(LLT::scalar(64), 8));
  297. auto &MF = *MI.getMF();
  298. auto *LowMMO = MF.getMachineMemOperand(&Store.getMMO(), 0, NewTy);
  299. auto *HighMMO = MF.getMachineMemOperand(&Store.getMMO(), 8, NewTy);
  300. B.buildStore(Zero, PtrReg, *LowMMO);
  301. B.buildStore(Zero, HighPtr, *HighMMO);
  302. Store.eraseFromParent();
  303. }
  304. #define AARCH64POSTLEGALIZERCOMBINERHELPER_GENCOMBINERHELPER_DEPS
  305. #include "AArch64GenPostLegalizeGICombiner.inc"
  306. #undef AARCH64POSTLEGALIZERCOMBINERHELPER_GENCOMBINERHELPER_DEPS
  307. namespace {
  308. #define AARCH64POSTLEGALIZERCOMBINERHELPER_GENCOMBINERHELPER_H
  309. #include "AArch64GenPostLegalizeGICombiner.inc"
  310. #undef AARCH64POSTLEGALIZERCOMBINERHELPER_GENCOMBINERHELPER_H
  311. class AArch64PostLegalizerCombinerInfo : public CombinerInfo {
  312. GISelKnownBits *KB;
  313. MachineDominatorTree *MDT;
  314. public:
  315. AArch64GenPostLegalizerCombinerHelperRuleConfig GeneratedRuleCfg;
  316. AArch64PostLegalizerCombinerInfo(bool EnableOpt, bool OptSize, bool MinSize,
  317. GISelKnownBits *KB,
  318. MachineDominatorTree *MDT)
  319. : CombinerInfo(/*AllowIllegalOps*/ true, /*ShouldLegalizeIllegal*/ false,
  320. /*LegalizerInfo*/ nullptr, EnableOpt, OptSize, MinSize),
  321. KB(KB), MDT(MDT) {
  322. if (!GeneratedRuleCfg.parseCommandLineOption())
  323. report_fatal_error("Invalid rule identifier");
  324. }
  325. bool combine(GISelChangeObserver &Observer, MachineInstr &MI,
  326. MachineIRBuilder &B) const override;
  327. };
  328. bool AArch64PostLegalizerCombinerInfo::combine(GISelChangeObserver &Observer,
  329. MachineInstr &MI,
  330. MachineIRBuilder &B) const {
  331. const auto *LI =
  332. MI.getParent()->getParent()->getSubtarget().getLegalizerInfo();
  333. CombinerHelper Helper(Observer, B, /*IsPreLegalize*/ false, KB, MDT, LI);
  334. AArch64GenPostLegalizerCombinerHelper Generated(GeneratedRuleCfg);
  335. return Generated.tryCombineAll(Observer, MI, B, Helper);
  336. }
  337. #define AARCH64POSTLEGALIZERCOMBINERHELPER_GENCOMBINERHELPER_CPP
  338. #include "AArch64GenPostLegalizeGICombiner.inc"
  339. #undef AARCH64POSTLEGALIZERCOMBINERHELPER_GENCOMBINERHELPER_CPP
  340. class AArch64PostLegalizerCombiner : public MachineFunctionPass {
  341. public:
  342. static char ID;
  343. AArch64PostLegalizerCombiner(bool IsOptNone = false);
  344. StringRef getPassName() const override {
  345. return "AArch64PostLegalizerCombiner";
  346. }
  347. bool runOnMachineFunction(MachineFunction &MF) override;
  348. void getAnalysisUsage(AnalysisUsage &AU) const override;
  349. private:
  350. bool IsOptNone;
  351. };
  352. } // end anonymous namespace
  353. void AArch64PostLegalizerCombiner::getAnalysisUsage(AnalysisUsage &AU) const {
  354. AU.addRequired<TargetPassConfig>();
  355. AU.setPreservesCFG();
  356. getSelectionDAGFallbackAnalysisUsage(AU);
  357. AU.addRequired<GISelKnownBitsAnalysis>();
  358. AU.addPreserved<GISelKnownBitsAnalysis>();
  359. if (!IsOptNone) {
  360. AU.addRequired<MachineDominatorTree>();
  361. AU.addPreserved<MachineDominatorTree>();
  362. AU.addRequired<GISelCSEAnalysisWrapperPass>();
  363. AU.addPreserved<GISelCSEAnalysisWrapperPass>();
  364. }
  365. MachineFunctionPass::getAnalysisUsage(AU);
  366. }
  367. AArch64PostLegalizerCombiner::AArch64PostLegalizerCombiner(bool IsOptNone)
  368. : MachineFunctionPass(ID), IsOptNone(IsOptNone) {
  369. initializeAArch64PostLegalizerCombinerPass(*PassRegistry::getPassRegistry());
  370. }
  371. bool AArch64PostLegalizerCombiner::runOnMachineFunction(MachineFunction &MF) {
  372. if (MF.getProperties().hasProperty(
  373. MachineFunctionProperties::Property::FailedISel))
  374. return false;
  375. assert(MF.getProperties().hasProperty(
  376. MachineFunctionProperties::Property::Legalized) &&
  377. "Expected a legalized function?");
  378. auto *TPC = &getAnalysis<TargetPassConfig>();
  379. const Function &F = MF.getFunction();
  380. bool EnableOpt =
  381. MF.getTarget().getOptLevel() != CodeGenOpt::None && !skipFunction(F);
  382. GISelKnownBits *KB = &getAnalysis<GISelKnownBitsAnalysis>().get(MF);
  383. MachineDominatorTree *MDT =
  384. IsOptNone ? nullptr : &getAnalysis<MachineDominatorTree>();
  385. AArch64PostLegalizerCombinerInfo PCInfo(EnableOpt, F.hasOptSize(),
  386. F.hasMinSize(), KB, MDT);
  387. GISelCSEAnalysisWrapper &Wrapper =
  388. getAnalysis<GISelCSEAnalysisWrapperPass>().getCSEWrapper();
  389. auto *CSEInfo = &Wrapper.get(TPC->getCSEConfig());
  390. Combiner C(PCInfo, TPC);
  391. return C.combineMachineInstrs(MF, CSEInfo);
  392. }
  393. char AArch64PostLegalizerCombiner::ID = 0;
  394. INITIALIZE_PASS_BEGIN(AArch64PostLegalizerCombiner, DEBUG_TYPE,
  395. "Combine AArch64 MachineInstrs after legalization", false,
  396. false)
  397. INITIALIZE_PASS_DEPENDENCY(TargetPassConfig)
  398. INITIALIZE_PASS_DEPENDENCY(GISelKnownBitsAnalysis)
  399. INITIALIZE_PASS_END(AArch64PostLegalizerCombiner, DEBUG_TYPE,
  400. "Combine AArch64 MachineInstrs after legalization", false,
  401. false)
  402. namespace llvm {
  403. FunctionPass *createAArch64PostLegalizerCombiner(bool IsOptNone) {
  404. return new AArch64PostLegalizerCombiner(IsOptNone);
  405. }
  406. } // end namespace llvm