123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763 |
- //===---- MachineCombiner.cpp - Instcombining on SSA form machine code ----===//
- //
- // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
- // See https://llvm.org/LICENSE.txt for license information.
- // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
- //
- //===----------------------------------------------------------------------===//
- //
- // The machine combiner pass uses machine trace metrics to ensure the combined
- // instructions do not lengthen the critical path or the resource depth.
- //===----------------------------------------------------------------------===//
- #include "llvm/ADT/DenseMap.h"
- #include "llvm/ADT/Statistic.h"
- #include "llvm/Analysis/ProfileSummaryInfo.h"
- #include "llvm/CodeGen/LazyMachineBlockFrequencyInfo.h"
- #include "llvm/CodeGen/MachineCombinerPattern.h"
- #include "llvm/CodeGen/MachineDominators.h"
- #include "llvm/CodeGen/MachineFunction.h"
- #include "llvm/CodeGen/MachineFunctionPass.h"
- #include "llvm/CodeGen/MachineLoopInfo.h"
- #include "llvm/CodeGen/MachineRegisterInfo.h"
- #include "llvm/CodeGen/MachineSizeOpts.h"
- #include "llvm/CodeGen/MachineTraceMetrics.h"
- #include "llvm/CodeGen/RegisterClassInfo.h"
- #include "llvm/CodeGen/TargetInstrInfo.h"
- #include "llvm/CodeGen/TargetRegisterInfo.h"
- #include "llvm/CodeGen/TargetSchedule.h"
- #include "llvm/CodeGen/TargetSubtargetInfo.h"
- #include "llvm/InitializePasses.h"
- #include "llvm/Support/CommandLine.h"
- #include "llvm/Support/Debug.h"
- #include "llvm/Support/raw_ostream.h"
- using namespace llvm;
- #define DEBUG_TYPE "machine-combiner"
- STATISTIC(NumInstCombined, "Number of machineinst combined");
- static cl::opt<unsigned>
- inc_threshold("machine-combiner-inc-threshold", cl::Hidden,
- cl::desc("Incremental depth computation will be used for basic "
- "blocks with more instructions."), cl::init(500));
- static cl::opt<bool> dump_intrs("machine-combiner-dump-subst-intrs", cl::Hidden,
- cl::desc("Dump all substituted intrs"),
- cl::init(false));
- #ifdef EXPENSIVE_CHECKS
- static cl::opt<bool> VerifyPatternOrder(
- "machine-combiner-verify-pattern-order", cl::Hidden,
- cl::desc(
- "Verify that the generated patterns are ordered by increasing latency"),
- cl::init(true));
- #else
- static cl::opt<bool> VerifyPatternOrder(
- "machine-combiner-verify-pattern-order", cl::Hidden,
- cl::desc(
- "Verify that the generated patterns are ordered by increasing latency"),
- cl::init(false));
- #endif
- namespace {
- class MachineCombiner : public MachineFunctionPass {
- const TargetSubtargetInfo *STI;
- const TargetInstrInfo *TII;
- const TargetRegisterInfo *TRI;
- MCSchedModel SchedModel;
- MachineRegisterInfo *MRI;
- MachineLoopInfo *MLI; // Current MachineLoopInfo
- MachineTraceMetrics *Traces;
- MachineTraceMetrics::Ensemble *MinInstr;
- MachineBlockFrequencyInfo *MBFI;
- ProfileSummaryInfo *PSI;
- RegisterClassInfo RegClassInfo;
- TargetSchedModel TSchedModel;
- /// True if optimizing for code size.
- bool OptSize;
- public:
- static char ID;
- MachineCombiner() : MachineFunctionPass(ID) {
- initializeMachineCombinerPass(*PassRegistry::getPassRegistry());
- }
- void getAnalysisUsage(AnalysisUsage &AU) const override;
- bool runOnMachineFunction(MachineFunction &MF) override;
- StringRef getPassName() const override { return "Machine InstCombiner"; }
- private:
- bool combineInstructions(MachineBasicBlock *);
- MachineInstr *getOperandDef(const MachineOperand &MO);
- bool isTransientMI(const MachineInstr *MI);
- unsigned getDepth(SmallVectorImpl<MachineInstr *> &InsInstrs,
- DenseMap<unsigned, unsigned> &InstrIdxForVirtReg,
- MachineTraceMetrics::Trace BlockTrace);
- unsigned getLatency(MachineInstr *Root, MachineInstr *NewRoot,
- MachineTraceMetrics::Trace BlockTrace);
- bool
- improvesCriticalPathLen(MachineBasicBlock *MBB, MachineInstr *Root,
- MachineTraceMetrics::Trace BlockTrace,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs,
- DenseMap<unsigned, unsigned> &InstrIdxForVirtReg,
- MachineCombinerPattern Pattern, bool SlackIsAccurate);
- bool reduceRegisterPressure(MachineInstr &Root, MachineBasicBlock *MBB,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs,
- MachineCombinerPattern Pattern);
- bool preservesResourceLen(MachineBasicBlock *MBB,
- MachineTraceMetrics::Trace BlockTrace,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs);
- void instr2instrSC(SmallVectorImpl<MachineInstr *> &Instrs,
- SmallVectorImpl<const MCSchedClassDesc *> &InstrsSC);
- std::pair<unsigned, unsigned>
- getLatenciesForInstrSequences(MachineInstr &MI,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs,
- MachineTraceMetrics::Trace BlockTrace);
- void verifyPatternOrder(MachineBasicBlock *MBB, MachineInstr &Root,
- SmallVector<MachineCombinerPattern, 16> &Patterns);
- };
- }
- char MachineCombiner::ID = 0;
- char &llvm::MachineCombinerID = MachineCombiner::ID;
- INITIALIZE_PASS_BEGIN(MachineCombiner, DEBUG_TYPE,
- "Machine InstCombiner", false, false)
- INITIALIZE_PASS_DEPENDENCY(MachineLoopInfo)
- INITIALIZE_PASS_DEPENDENCY(MachineTraceMetrics)
- INITIALIZE_PASS_END(MachineCombiner, DEBUG_TYPE, "Machine InstCombiner",
- false, false)
- void MachineCombiner::getAnalysisUsage(AnalysisUsage &AU) const {
- AU.setPreservesCFG();
- AU.addPreserved<MachineDominatorTree>();
- AU.addRequired<MachineLoopInfo>();
- AU.addPreserved<MachineLoopInfo>();
- AU.addRequired<MachineTraceMetrics>();
- AU.addPreserved<MachineTraceMetrics>();
- AU.addRequired<LazyMachineBlockFrequencyInfoPass>();
- AU.addRequired<ProfileSummaryInfoWrapperPass>();
- MachineFunctionPass::getAnalysisUsage(AU);
- }
- MachineInstr *MachineCombiner::getOperandDef(const MachineOperand &MO) {
- MachineInstr *DefInstr = nullptr;
- // We need a virtual register definition.
- if (MO.isReg() && MO.getReg().isVirtual())
- DefInstr = MRI->getUniqueVRegDef(MO.getReg());
- // PHI's have no depth etc.
- if (DefInstr && DefInstr->isPHI())
- DefInstr = nullptr;
- return DefInstr;
- }
- /// Return true if MI is unlikely to generate an actual target instruction.
- bool MachineCombiner::isTransientMI(const MachineInstr *MI) {
- if (!MI->isCopy())
- return MI->isTransient();
- // If MI is a COPY, check if its src and dst registers can be coalesced.
- Register Dst = MI->getOperand(0).getReg();
- Register Src = MI->getOperand(1).getReg();
- if (!MI->isFullCopy()) {
- // If src RC contains super registers of dst RC, it can also be coalesced.
- if (MI->getOperand(0).getSubReg() || Src.isPhysical() || Dst.isPhysical())
- return false;
- auto SrcSub = MI->getOperand(1).getSubReg();
- auto SrcRC = MRI->getRegClass(Src);
- auto DstRC = MRI->getRegClass(Dst);
- return TRI->getMatchingSuperRegClass(SrcRC, DstRC, SrcSub) != nullptr;
- }
- if (Src.isPhysical() && Dst.isPhysical())
- return Src == Dst;
- if (Src.isVirtual() && Dst.isVirtual()) {
- auto SrcRC = MRI->getRegClass(Src);
- auto DstRC = MRI->getRegClass(Dst);
- return SrcRC->hasSuperClassEq(DstRC) || SrcRC->hasSubClassEq(DstRC);
- }
- if (Src.isVirtual())
- std::swap(Src, Dst);
- // Now Src is physical register, Dst is virtual register.
- auto DstRC = MRI->getRegClass(Dst);
- return DstRC->contains(Src);
- }
- /// Computes depth of instructions in vector \InsInstr.
- ///
- /// \param InsInstrs is a vector of machine instructions
- /// \param InstrIdxForVirtReg is a dense map of virtual register to index
- /// of defining machine instruction in \p InsInstrs
- /// \param BlockTrace is a trace of machine instructions
- ///
- /// \returns Depth of last instruction in \InsInstrs ("NewRoot")
- unsigned
- MachineCombiner::getDepth(SmallVectorImpl<MachineInstr *> &InsInstrs,
- DenseMap<unsigned, unsigned> &InstrIdxForVirtReg,
- MachineTraceMetrics::Trace BlockTrace) {
- SmallVector<unsigned, 16> InstrDepth;
- // For each instruction in the new sequence compute the depth based on the
- // operands. Use the trace information when possible. For new operands which
- // are tracked in the InstrIdxForVirtReg map depth is looked up in InstrDepth
- for (auto *InstrPtr : InsInstrs) { // for each Use
- unsigned IDepth = 0;
- for (const MachineOperand &MO : InstrPtr->operands()) {
- // Check for virtual register operand.
- if (!(MO.isReg() && MO.getReg().isVirtual()))
- continue;
- if (!MO.isUse())
- continue;
- unsigned DepthOp = 0;
- unsigned LatencyOp = 0;
- DenseMap<unsigned, unsigned>::iterator II =
- InstrIdxForVirtReg.find(MO.getReg());
- if (II != InstrIdxForVirtReg.end()) {
- // Operand is new virtual register not in trace
- assert(II->second < InstrDepth.size() && "Bad Index");
- MachineInstr *DefInstr = InsInstrs[II->second];
- assert(DefInstr &&
- "There must be a definition for a new virtual register");
- DepthOp = InstrDepth[II->second];
- int DefIdx = DefInstr->findRegisterDefOperandIdx(MO.getReg());
- int UseIdx = InstrPtr->findRegisterUseOperandIdx(MO.getReg());
- LatencyOp = TSchedModel.computeOperandLatency(DefInstr, DefIdx,
- InstrPtr, UseIdx);
- } else {
- MachineInstr *DefInstr = getOperandDef(MO);
- if (DefInstr) {
- DepthOp = BlockTrace.getInstrCycles(*DefInstr).Depth;
- if (!isTransientMI(DefInstr))
- LatencyOp = TSchedModel.computeOperandLatency(
- DefInstr, DefInstr->findRegisterDefOperandIdx(MO.getReg()),
- InstrPtr, InstrPtr->findRegisterUseOperandIdx(MO.getReg()));
- }
- }
- IDepth = std::max(IDepth, DepthOp + LatencyOp);
- }
- InstrDepth.push_back(IDepth);
- }
- unsigned NewRootIdx = InsInstrs.size() - 1;
- return InstrDepth[NewRootIdx];
- }
- /// Computes instruction latency as max of latency of defined operands.
- ///
- /// \param Root is a machine instruction that could be replaced by NewRoot.
- /// It is used to compute a more accurate latency information for NewRoot in
- /// case there is a dependent instruction in the same trace (\p BlockTrace)
- /// \param NewRoot is the instruction for which the latency is computed
- /// \param BlockTrace is a trace of machine instructions
- ///
- /// \returns Latency of \p NewRoot
- unsigned MachineCombiner::getLatency(MachineInstr *Root, MachineInstr *NewRoot,
- MachineTraceMetrics::Trace BlockTrace) {
- // Check each definition in NewRoot and compute the latency
- unsigned NewRootLatency = 0;
- for (const MachineOperand &MO : NewRoot->operands()) {
- // Check for virtual register operand.
- if (!(MO.isReg() && MO.getReg().isVirtual()))
- continue;
- if (!MO.isDef())
- continue;
- // Get the first instruction that uses MO
- MachineRegisterInfo::reg_iterator RI = MRI->reg_begin(MO.getReg());
- RI++;
- if (RI == MRI->reg_end())
- continue;
- MachineInstr *UseMO = RI->getParent();
- unsigned LatencyOp = 0;
- if (UseMO && BlockTrace.isDepInTrace(*Root, *UseMO)) {
- LatencyOp = TSchedModel.computeOperandLatency(
- NewRoot, NewRoot->findRegisterDefOperandIdx(MO.getReg()), UseMO,
- UseMO->findRegisterUseOperandIdx(MO.getReg()));
- } else {
- LatencyOp = TSchedModel.computeInstrLatency(NewRoot);
- }
- NewRootLatency = std::max(NewRootLatency, LatencyOp);
- }
- return NewRootLatency;
- }
- /// The combiner's goal may differ based on which pattern it is attempting
- /// to optimize.
- enum class CombinerObjective {
- MustReduceDepth, // The data dependency chain must be improved.
- MustReduceRegisterPressure, // The register pressure must be reduced.
- Default // The critical path must not be lengthened.
- };
- static CombinerObjective getCombinerObjective(MachineCombinerPattern P) {
- // TODO: If C++ ever gets a real enum class, make this part of the
- // MachineCombinerPattern class.
- switch (P) {
- case MachineCombinerPattern::REASSOC_AX_BY:
- case MachineCombinerPattern::REASSOC_AX_YB:
- case MachineCombinerPattern::REASSOC_XA_BY:
- case MachineCombinerPattern::REASSOC_XA_YB:
- case MachineCombinerPattern::REASSOC_XY_AMM_BMM:
- case MachineCombinerPattern::REASSOC_XMM_AMM_BMM:
- case MachineCombinerPattern::SUBADD_OP1:
- case MachineCombinerPattern::SUBADD_OP2:
- case MachineCombinerPattern::FMADD_AX:
- case MachineCombinerPattern::FMADD_XA:
- case MachineCombinerPattern::FMSUB:
- case MachineCombinerPattern::FNMSUB:
- return CombinerObjective::MustReduceDepth;
- case MachineCombinerPattern::REASSOC_XY_BCA:
- case MachineCombinerPattern::REASSOC_XY_BAC:
- return CombinerObjective::MustReduceRegisterPressure;
- default:
- return CombinerObjective::Default;
- }
- }
- /// Estimate the latency of the new and original instruction sequence by summing
- /// up the latencies of the inserted and deleted instructions. This assumes
- /// that the inserted and deleted instructions are dependent instruction chains,
- /// which might not hold in all cases.
- std::pair<unsigned, unsigned> MachineCombiner::getLatenciesForInstrSequences(
- MachineInstr &MI, SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs,
- MachineTraceMetrics::Trace BlockTrace) {
- assert(!InsInstrs.empty() && "Only support sequences that insert instrs.");
- unsigned NewRootLatency = 0;
- // NewRoot is the last instruction in the \p InsInstrs vector.
- MachineInstr *NewRoot = InsInstrs.back();
- for (unsigned i = 0; i < InsInstrs.size() - 1; i++)
- NewRootLatency += TSchedModel.computeInstrLatency(InsInstrs[i]);
- NewRootLatency += getLatency(&MI, NewRoot, BlockTrace);
- unsigned RootLatency = 0;
- for (auto *I : DelInstrs)
- RootLatency += TSchedModel.computeInstrLatency(I);
- return {NewRootLatency, RootLatency};
- }
- bool MachineCombiner::reduceRegisterPressure(
- MachineInstr &Root, MachineBasicBlock *MBB,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs,
- MachineCombinerPattern Pattern) {
- // FIXME: for now, we don't do any check for the register pressure patterns.
- // We treat them as always profitable. But we can do better if we make
- // RegPressureTracker class be aware of TIE attribute. Then we can get an
- // accurate compare of register pressure with DelInstrs or InsInstrs.
- return true;
- }
- /// The DAGCombine code sequence ends in MI (Machine Instruction) Root.
- /// The new code sequence ends in MI NewRoot. A necessary condition for the new
- /// sequence to replace the old sequence is that it cannot lengthen the critical
- /// path. The definition of "improve" may be restricted by specifying that the
- /// new path improves the data dependency chain (MustReduceDepth).
- bool MachineCombiner::improvesCriticalPathLen(
- MachineBasicBlock *MBB, MachineInstr *Root,
- MachineTraceMetrics::Trace BlockTrace,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs,
- DenseMap<unsigned, unsigned> &InstrIdxForVirtReg,
- MachineCombinerPattern Pattern,
- bool SlackIsAccurate) {
- // Get depth and latency of NewRoot and Root.
- unsigned NewRootDepth = getDepth(InsInstrs, InstrIdxForVirtReg, BlockTrace);
- unsigned RootDepth = BlockTrace.getInstrCycles(*Root).Depth;
- LLVM_DEBUG(dbgs() << " Dependence data for " << *Root << "\tNewRootDepth: "
- << NewRootDepth << "\tRootDepth: " << RootDepth);
- // For a transform such as reassociation, the cost equation is
- // conservatively calculated so that we must improve the depth (data
- // dependency cycles) in the critical path to proceed with the transform.
- // Being conservative also protects against inaccuracies in the underlying
- // machine trace metrics and CPU models.
- if (getCombinerObjective(Pattern) == CombinerObjective::MustReduceDepth) {
- LLVM_DEBUG(dbgs() << "\tIt MustReduceDepth ");
- LLVM_DEBUG(NewRootDepth < RootDepth
- ? dbgs() << "\t and it does it\n"
- : dbgs() << "\t but it does NOT do it\n");
- return NewRootDepth < RootDepth;
- }
- // A more flexible cost calculation for the critical path includes the slack
- // of the original code sequence. This may allow the transform to proceed
- // even if the instruction depths (data dependency cycles) become worse.
- // Account for the latency of the inserted and deleted instructions by
- unsigned NewRootLatency, RootLatency;
- std::tie(NewRootLatency, RootLatency) =
- getLatenciesForInstrSequences(*Root, InsInstrs, DelInstrs, BlockTrace);
- unsigned RootSlack = BlockTrace.getInstrSlack(*Root);
- unsigned NewCycleCount = NewRootDepth + NewRootLatency;
- unsigned OldCycleCount =
- RootDepth + RootLatency + (SlackIsAccurate ? RootSlack : 0);
- LLVM_DEBUG(dbgs() << "\n\tNewRootLatency: " << NewRootLatency
- << "\tRootLatency: " << RootLatency << "\n\tRootSlack: "
- << RootSlack << " SlackIsAccurate=" << SlackIsAccurate
- << "\n\tNewRootDepth + NewRootLatency = " << NewCycleCount
- << "\n\tRootDepth + RootLatency + RootSlack = "
- << OldCycleCount;);
- LLVM_DEBUG(NewCycleCount <= OldCycleCount
- ? dbgs() << "\n\t It IMPROVES PathLen because"
- : dbgs() << "\n\t It DOES NOT improve PathLen because");
- LLVM_DEBUG(dbgs() << "\n\t\tNewCycleCount = " << NewCycleCount
- << ", OldCycleCount = " << OldCycleCount << "\n");
- return NewCycleCount <= OldCycleCount;
- }
- /// helper routine to convert instructions into SC
- void MachineCombiner::instr2instrSC(
- SmallVectorImpl<MachineInstr *> &Instrs,
- SmallVectorImpl<const MCSchedClassDesc *> &InstrsSC) {
- for (auto *InstrPtr : Instrs) {
- unsigned Opc = InstrPtr->getOpcode();
- unsigned Idx = TII->get(Opc).getSchedClass();
- const MCSchedClassDesc *SC = SchedModel.getSchedClassDesc(Idx);
- InstrsSC.push_back(SC);
- }
- }
- /// True when the new instructions do not increase resource length
- bool MachineCombiner::preservesResourceLen(
- MachineBasicBlock *MBB, MachineTraceMetrics::Trace BlockTrace,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs) {
- if (!TSchedModel.hasInstrSchedModel())
- return true;
- // Compute current resource length
- //ArrayRef<const MachineBasicBlock *> MBBarr(MBB);
- SmallVector <const MachineBasicBlock *, 1> MBBarr;
- MBBarr.push_back(MBB);
- unsigned ResLenBeforeCombine = BlockTrace.getResourceLength(MBBarr);
- // Deal with SC rather than Instructions.
- SmallVector<const MCSchedClassDesc *, 16> InsInstrsSC;
- SmallVector<const MCSchedClassDesc *, 16> DelInstrsSC;
- instr2instrSC(InsInstrs, InsInstrsSC);
- instr2instrSC(DelInstrs, DelInstrsSC);
- ArrayRef<const MCSchedClassDesc *> MSCInsArr{InsInstrsSC};
- ArrayRef<const MCSchedClassDesc *> MSCDelArr{DelInstrsSC};
- // Compute new resource length.
- unsigned ResLenAfterCombine =
- BlockTrace.getResourceLength(MBBarr, MSCInsArr, MSCDelArr);
- LLVM_DEBUG(dbgs() << "\t\tResource length before replacement: "
- << ResLenBeforeCombine
- << " and after: " << ResLenAfterCombine << "\n";);
- LLVM_DEBUG(
- ResLenAfterCombine <=
- ResLenBeforeCombine + TII->getExtendResourceLenLimit()
- ? dbgs() << "\t\t As result it IMPROVES/PRESERVES Resource Length\n"
- : dbgs() << "\t\t As result it DOES NOT improve/preserve Resource "
- "Length\n");
- return ResLenAfterCombine <=
- ResLenBeforeCombine + TII->getExtendResourceLenLimit();
- }
- /// Inserts InsInstrs and deletes DelInstrs. Incrementally updates instruction
- /// depths if requested.
- ///
- /// \param MBB basic block to insert instructions in
- /// \param MI current machine instruction
- /// \param InsInstrs new instructions to insert in \p MBB
- /// \param DelInstrs instruction to delete from \p MBB
- /// \param MinInstr is a pointer to the machine trace information
- /// \param RegUnits set of live registers, needed to compute instruction depths
- /// \param TII is target instruction info, used to call target hook
- /// \param Pattern is used to call target hook finalizeInsInstrs
- /// \param IncrementalUpdate if true, compute instruction depths incrementally,
- /// otherwise invalidate the trace
- static void insertDeleteInstructions(MachineBasicBlock *MBB, MachineInstr &MI,
- SmallVector<MachineInstr *, 16> InsInstrs,
- SmallVector<MachineInstr *, 16> DelInstrs,
- MachineTraceMetrics::Ensemble *MinInstr,
- SparseSet<LiveRegUnit> &RegUnits,
- const TargetInstrInfo *TII,
- MachineCombinerPattern Pattern,
- bool IncrementalUpdate) {
- // If we want to fix up some placeholder for some target, do it now.
- // We need this because in genAlternativeCodeSequence, we have not decided the
- // better pattern InsInstrs or DelInstrs, so we don't want generate some
- // sideeffect to the function. For example we need to delay the constant pool
- // entry creation here after InsInstrs is selected as better pattern.
- // Otherwise the constant pool entry created for InsInstrs will not be deleted
- // even if InsInstrs is not the better pattern.
- TII->finalizeInsInstrs(MI, Pattern, InsInstrs);
- for (auto *InstrPtr : InsInstrs)
- MBB->insert((MachineBasicBlock::iterator)&MI, InstrPtr);
- for (auto *InstrPtr : DelInstrs) {
- InstrPtr->eraseFromParent();
- // Erase all LiveRegs defined by the removed instruction
- for (auto *I = RegUnits.begin(); I != RegUnits.end();) {
- if (I->MI == InstrPtr)
- I = RegUnits.erase(I);
- else
- I++;
- }
- }
- if (IncrementalUpdate)
- for (auto *InstrPtr : InsInstrs)
- MinInstr->updateDepth(MBB, *InstrPtr, RegUnits);
- else
- MinInstr->invalidate(MBB);
- NumInstCombined++;
- }
- // Check that the difference between original and new latency is decreasing for
- // later patterns. This helps to discover sub-optimal pattern orderings.
- void MachineCombiner::verifyPatternOrder(
- MachineBasicBlock *MBB, MachineInstr &Root,
- SmallVector<MachineCombinerPattern, 16> &Patterns) {
- long PrevLatencyDiff = std::numeric_limits<long>::max();
- (void)PrevLatencyDiff; // Variable is used in assert only.
- for (auto P : Patterns) {
- SmallVector<MachineInstr *, 16> InsInstrs;
- SmallVector<MachineInstr *, 16> DelInstrs;
- DenseMap<unsigned, unsigned> InstrIdxForVirtReg;
- TII->genAlternativeCodeSequence(Root, P, InsInstrs, DelInstrs,
- InstrIdxForVirtReg);
- // Found pattern, but did not generate alternative sequence.
- // This can happen e.g. when an immediate could not be materialized
- // in a single instruction.
- if (InsInstrs.empty() || !TSchedModel.hasInstrSchedModelOrItineraries())
- continue;
- unsigned NewRootLatency, RootLatency;
- std::tie(NewRootLatency, RootLatency) = getLatenciesForInstrSequences(
- Root, InsInstrs, DelInstrs, MinInstr->getTrace(MBB));
- long CurrentLatencyDiff = ((long)RootLatency) - ((long)NewRootLatency);
- assert(CurrentLatencyDiff <= PrevLatencyDiff &&
- "Current pattern is better than previous pattern.");
- PrevLatencyDiff = CurrentLatencyDiff;
- }
- }
- /// Substitute a slow code sequence with a faster one by
- /// evaluating instruction combining pattern.
- /// The prototype of such a pattern is MUl + ADD -> MADD. Performs instruction
- /// combining based on machine trace metrics. Only combine a sequence of
- /// instructions when this neither lengthens the critical path nor increases
- /// resource pressure. When optimizing for codesize always combine when the new
- /// sequence is shorter.
- bool MachineCombiner::combineInstructions(MachineBasicBlock *MBB) {
- bool Changed = false;
- LLVM_DEBUG(dbgs() << "Combining MBB " << MBB->getName() << "\n");
- bool IncrementalUpdate = false;
- auto BlockIter = MBB->begin();
- decltype(BlockIter) LastUpdate;
- // Check if the block is in a loop.
- const MachineLoop *ML = MLI->getLoopFor(MBB);
- if (!MinInstr)
- MinInstr = Traces->getEnsemble(MachineTraceMetrics::TS_MinInstrCount);
- SparseSet<LiveRegUnit> RegUnits;
- RegUnits.setUniverse(TRI->getNumRegUnits());
- bool OptForSize = OptSize || llvm::shouldOptimizeForSize(MBB, PSI, MBFI);
- bool DoRegPressureReduce =
- TII->shouldReduceRegisterPressure(MBB, &RegClassInfo);
- while (BlockIter != MBB->end()) {
- auto &MI = *BlockIter++;
- SmallVector<MachineCombinerPattern, 16> Patterns;
- // The motivating example is:
- //
- // MUL Other MUL_op1 MUL_op2 Other
- // \ / \ | /
- // ADD/SUB => MADD/MSUB
- // (=Root) (=NewRoot)
- // The DAGCombine code always replaced MUL + ADD/SUB by MADD. While this is
- // usually beneficial for code size it unfortunately can hurt performance
- // when the ADD is on the critical path, but the MUL is not. With the
- // substitution the MUL becomes part of the critical path (in form of the
- // MADD) and can lengthen it on architectures where the MADD latency is
- // longer than the ADD latency.
- //
- // For each instruction we check if it can be the root of a combiner
- // pattern. Then for each pattern the new code sequence in form of MI is
- // generated and evaluated. When the efficiency criteria (don't lengthen
- // critical path, don't use more resources) is met the new sequence gets
- // hooked up into the basic block before the old sequence is removed.
- //
- // The algorithm does not try to evaluate all patterns and pick the best.
- // This is only an artificial restriction though. In practice there is
- // mostly one pattern, and getMachineCombinerPatterns() can order patterns
- // based on an internal cost heuristic. If
- // machine-combiner-verify-pattern-order is enabled, all patterns are
- // checked to ensure later patterns do not provide better latency savings.
- if (!TII->getMachineCombinerPatterns(MI, Patterns, DoRegPressureReduce))
- continue;
- if (VerifyPatternOrder)
- verifyPatternOrder(MBB, MI, Patterns);
- for (const auto P : Patterns) {
- SmallVector<MachineInstr *, 16> InsInstrs;
- SmallVector<MachineInstr *, 16> DelInstrs;
- DenseMap<unsigned, unsigned> InstrIdxForVirtReg;
- TII->genAlternativeCodeSequence(MI, P, InsInstrs, DelInstrs,
- InstrIdxForVirtReg);
- // Found pattern, but did not generate alternative sequence.
- // This can happen e.g. when an immediate could not be materialized
- // in a single instruction.
- if (InsInstrs.empty())
- continue;
- LLVM_DEBUG(if (dump_intrs) {
- dbgs() << "\tFor the Pattern (" << (int)P
- << ") these instructions could be removed\n";
- for (auto const *InstrPtr : DelInstrs)
- InstrPtr->print(dbgs(), /*IsStandalone*/false, /*SkipOpers*/false,
- /*SkipDebugLoc*/false, /*AddNewLine*/true, TII);
- dbgs() << "\tThese instructions could replace the removed ones\n";
- for (auto const *InstrPtr : InsInstrs)
- InstrPtr->print(dbgs(), /*IsStandalone*/false, /*SkipOpers*/false,
- /*SkipDebugLoc*/false, /*AddNewLine*/true, TII);
- });
- if (IncrementalUpdate && LastUpdate != BlockIter) {
- // Update depths since the last incremental update.
- MinInstr->updateDepths(LastUpdate, BlockIter, RegUnits);
- LastUpdate = BlockIter;
- }
- if (DoRegPressureReduce &&
- getCombinerObjective(P) ==
- CombinerObjective::MustReduceRegisterPressure) {
- if (MBB->size() > inc_threshold) {
- // Use incremental depth updates for basic blocks above threshold
- IncrementalUpdate = true;
- LastUpdate = BlockIter;
- }
- if (reduceRegisterPressure(MI, MBB, InsInstrs, DelInstrs, P)) {
- // Replace DelInstrs with InsInstrs.
- insertDeleteInstructions(MBB, MI, InsInstrs, DelInstrs, MinInstr,
- RegUnits, TII, P, IncrementalUpdate);
- Changed |= true;
- // Go back to previous instruction as it may have ILP reassociation
- // opportunity.
- BlockIter--;
- break;
- }
- }
- if (ML && TII->isThroughputPattern(P)) {
- LLVM_DEBUG(dbgs() << "\t Replacing due to throughput pattern in loop\n");
- insertDeleteInstructions(MBB, MI, InsInstrs, DelInstrs, MinInstr,
- RegUnits, TII, P, IncrementalUpdate);
- // Eagerly stop after the first pattern fires.
- Changed = true;
- break;
- } else if (OptForSize && InsInstrs.size() < DelInstrs.size()) {
- LLVM_DEBUG(dbgs() << "\t Replacing due to OptForSize ("
- << InsInstrs.size() << " < "
- << DelInstrs.size() << ")\n");
- insertDeleteInstructions(MBB, MI, InsInstrs, DelInstrs, MinInstr,
- RegUnits, TII, P, IncrementalUpdate);
- // Eagerly stop after the first pattern fires.
- Changed = true;
- break;
- } else {
- // For big basic blocks, we only compute the full trace the first time
- // we hit this. We do not invalidate the trace, but instead update the
- // instruction depths incrementally.
- // NOTE: Only the instruction depths up to MI are accurate. All other
- // trace information is not updated.
- MachineTraceMetrics::Trace BlockTrace = MinInstr->getTrace(MBB);
- Traces->verifyAnalysis();
- if (improvesCriticalPathLen(MBB, &MI, BlockTrace, InsInstrs, DelInstrs,
- InstrIdxForVirtReg, P,
- !IncrementalUpdate) &&
- preservesResourceLen(MBB, BlockTrace, InsInstrs, DelInstrs)) {
- if (MBB->size() > inc_threshold) {
- // Use incremental depth updates for basic blocks above treshold
- IncrementalUpdate = true;
- LastUpdate = BlockIter;
- }
- insertDeleteInstructions(MBB, MI, InsInstrs, DelInstrs, MinInstr,
- RegUnits, TII, P, IncrementalUpdate);
- // Eagerly stop after the first pattern fires.
- Changed = true;
- break;
- }
- // Cleanup instructions of the alternative code sequence. There is no
- // use for them.
- MachineFunction *MF = MBB->getParent();
- for (auto *InstrPtr : InsInstrs)
- MF->deleteMachineInstr(InstrPtr);
- }
- InstrIdxForVirtReg.clear();
- }
- }
- if (Changed && IncrementalUpdate)
- Traces->invalidate(MBB);
- return Changed;
- }
- bool MachineCombiner::runOnMachineFunction(MachineFunction &MF) {
- STI = &MF.getSubtarget();
- TII = STI->getInstrInfo();
- TRI = STI->getRegisterInfo();
- SchedModel = STI->getSchedModel();
- TSchedModel.init(STI);
- MRI = &MF.getRegInfo();
- MLI = &getAnalysis<MachineLoopInfo>();
- Traces = &getAnalysis<MachineTraceMetrics>();
- PSI = &getAnalysis<ProfileSummaryInfoWrapperPass>().getPSI();
- MBFI = (PSI && PSI->hasProfileSummary()) ?
- &getAnalysis<LazyMachineBlockFrequencyInfoPass>().getBFI() :
- nullptr;
- MinInstr = nullptr;
- OptSize = MF.getFunction().hasOptSize();
- RegClassInfo.runOnMachineFunction(MF);
- LLVM_DEBUG(dbgs() << getPassName() << ": " << MF.getName() << '\n');
- if (!TII->useMachineCombiner()) {
- LLVM_DEBUG(
- dbgs()
- << " Skipping pass: Target does not support machine combiner\n");
- return false;
- }
- bool Changed = false;
- // Try to combine instructions.
- for (auto &MBB : MF)
- Changed |= combineInstructions(&MBB);
- return Changed;
- }
|