123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707 |
- //===-- PPCInstrInfo.h - PowerPC Instruction Information --------*- C++ -*-===//
- //
- // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
- // See https://llvm.org/LICENSE.txt for license information.
- // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
- //
- //===----------------------------------------------------------------------===//
- //
- // This file contains the PowerPC implementation of the TargetInstrInfo class.
- //
- //===----------------------------------------------------------------------===//
- #ifndef LLVM_LIB_TARGET_POWERPC_PPCINSTRINFO_H
- #define LLVM_LIB_TARGET_POWERPC_PPCINSTRINFO_H
- #include "PPCRegisterInfo.h"
- #include "llvm/CodeGen/TargetInstrInfo.h"
- #define GET_INSTRINFO_HEADER
- #include "PPCGenInstrInfo.inc"
- namespace llvm {
- /// PPCII - This namespace holds all of the PowerPC target-specific
- /// per-instruction flags. These must match the corresponding definitions in
- /// PPC.td and PPCInstrFormats.td.
- namespace PPCII {
- enum {
- // PPC970 Instruction Flags. These flags describe the characteristics of the
- // PowerPC 970 (aka G5) dispatch groups and how they are formed out of
- // raw machine instructions.
- /// PPC970_First - This instruction starts a new dispatch group, so it will
- /// always be the first one in the group.
- PPC970_First = 0x1,
- /// PPC970_Single - This instruction starts a new dispatch group and
- /// terminates it, so it will be the sole instruction in the group.
- PPC970_Single = 0x2,
- /// PPC970_Cracked - This instruction is cracked into two pieces, requiring
- /// two dispatch pipes to be available to issue.
- PPC970_Cracked = 0x4,
- /// PPC970_Mask/Shift - This is a bitmask that selects the pipeline type that
- /// an instruction is issued to.
- PPC970_Shift = 3,
- PPC970_Mask = 0x07 << PPC970_Shift
- };
- enum PPC970_Unit {
- /// These are the various PPC970 execution unit pipelines. Each instruction
- /// is one of these.
- PPC970_Pseudo = 0 << PPC970_Shift, // Pseudo instruction
- PPC970_FXU = 1 << PPC970_Shift, // Fixed Point (aka Integer/ALU) Unit
- PPC970_LSU = 2 << PPC970_Shift, // Load Store Unit
- PPC970_FPU = 3 << PPC970_Shift, // Floating Point Unit
- PPC970_CRU = 4 << PPC970_Shift, // Control Register Unit
- PPC970_VALU = 5 << PPC970_Shift, // Vector ALU
- PPC970_VPERM = 6 << PPC970_Shift, // Vector Permute Unit
- PPC970_BRU = 7 << PPC970_Shift // Branch Unit
- };
- enum {
- /// Shift count to bypass PPC970 flags
- NewDef_Shift = 6,
- /// This instruction is an X-Form memory operation.
- XFormMemOp = 0x1 << NewDef_Shift,
- /// This instruction is prefixed.
- Prefixed = 0x1 << (NewDef_Shift+1)
- };
- } // end namespace PPCII
- // Instructions that have an immediate form might be convertible to that
- // form if the correct input is a result of a load immediate. In order to
- // know whether the transformation is special, we might need to know some
- // of the details of the two forms.
- struct ImmInstrInfo {
- // Is the immediate field in the immediate form signed or unsigned?
- uint64_t SignedImm : 1;
- // Does the immediate need to be a multiple of some value?
- uint64_t ImmMustBeMultipleOf : 5;
- // Is R0/X0 treated specially by the original r+r instruction?
- // If so, in which operand?
- uint64_t ZeroIsSpecialOrig : 3;
- // Is R0/X0 treated specially by the new r+i instruction?
- // If so, in which operand?
- uint64_t ZeroIsSpecialNew : 3;
- // Is the operation commutative?
- uint64_t IsCommutative : 1;
- // The operand number to check for add-immediate def.
- uint64_t OpNoForForwarding : 3;
- // The operand number for the immediate.
- uint64_t ImmOpNo : 3;
- // The opcode of the new instruction.
- uint64_t ImmOpcode : 16;
- // The size of the immediate.
- uint64_t ImmWidth : 5;
- // The immediate should be truncated to N bits.
- uint64_t TruncateImmTo : 5;
- // Is the instruction summing the operand
- uint64_t IsSummingOperands : 1;
- };
- // Information required to convert an instruction to just a materialized
- // immediate.
- struct LoadImmediateInfo {
- unsigned Imm : 16;
- unsigned Is64Bit : 1;
- unsigned SetCR : 1;
- };
- // Index into the OpcodesForSpill array.
- enum SpillOpcodeKey {
- SOK_Int4Spill,
- SOK_Int8Spill,
- SOK_Float8Spill,
- SOK_Float4Spill,
- SOK_CRSpill,
- SOK_CRBitSpill,
- SOK_VRVectorSpill,
- SOK_VSXVectorSpill,
- SOK_VectorFloat8Spill,
- SOK_VectorFloat4Spill,
- SOK_SpillToVSR,
- SOK_PairedVecSpill,
- SOK_AccumulatorSpill,
- SOK_UAccumulatorSpill,
- SOK_SPESpill,
- SOK_PairedG8Spill,
- SOK_LastOpcodeSpill // This must be last on the enum.
- };
- // Define list of load and store spill opcodes.
- #define NoInstr PPC::INSTRUCTION_LIST_END
- #define Pwr8LoadOpcodes \
- { \
- PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR, \
- PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXVD2X, PPC::LXSDX, PPC::LXSSPX, \
- PPC::SPILLTOVSR_LD, NoInstr, NoInstr, NoInstr, PPC::EVLDD, \
- PPC::RESTORE_QUADWORD \
- }
- #define Pwr9LoadOpcodes \
- { \
- PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR, \
- PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXV, PPC::DFLOADf64, \
- PPC::DFLOADf32, PPC::SPILLTOVSR_LD, NoInstr, NoInstr, NoInstr, \
- NoInstr, PPC::RESTORE_QUADWORD \
- }
- #define Pwr10LoadOpcodes \
- { \
- PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR, \
- PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXV, PPC::DFLOADf64, \
- PPC::DFLOADf32, PPC::SPILLTOVSR_LD, PPC::LXVP, PPC::RESTORE_ACC, \
- PPC::RESTORE_UACC, NoInstr, PPC::RESTORE_QUADWORD \
- }
- #define Pwr8StoreOpcodes \
- { \
- PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \
- PPC::STVX, PPC::STXVD2X, PPC::STXSDX, PPC::STXSSPX, \
- PPC::SPILLTOVSR_ST, NoInstr, NoInstr, NoInstr, PPC::EVSTDD, \
- PPC::SPILL_QUADWORD \
- }
- #define Pwr9StoreOpcodes \
- { \
- PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \
- PPC::STVX, PPC::STXV, PPC::DFSTOREf64, PPC::DFSTOREf32, \
- PPC::SPILLTOVSR_ST, NoInstr, NoInstr, NoInstr, NoInstr, \
- PPC::SPILL_QUADWORD \
- }
- #define Pwr10StoreOpcodes \
- { \
- PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \
- PPC::STVX, PPC::STXV, PPC::DFSTOREf64, PPC::DFSTOREf32, \
- PPC::SPILLTOVSR_ST, PPC::STXVP, PPC::SPILL_ACC, PPC::SPILL_UACC, \
- NoInstr, PPC::SPILL_QUADWORD \
- }
- // Initialize arrays for load and store spill opcodes on supported subtargets.
- #define StoreOpcodesForSpill \
- { Pwr8StoreOpcodes, Pwr9StoreOpcodes, Pwr10StoreOpcodes }
- #define LoadOpcodesForSpill \
- { Pwr8LoadOpcodes, Pwr9LoadOpcodes, Pwr10LoadOpcodes }
- class PPCSubtarget;
- class PPCInstrInfo : public PPCGenInstrInfo {
- PPCSubtarget &Subtarget;
- const PPCRegisterInfo RI;
- const unsigned StoreSpillOpcodesArray[3][SOK_LastOpcodeSpill] =
- StoreOpcodesForSpill;
- const unsigned LoadSpillOpcodesArray[3][SOK_LastOpcodeSpill] =
- LoadOpcodesForSpill;
- void StoreRegToStackSlot(MachineFunction &MF, unsigned SrcReg, bool isKill,
- int FrameIdx, const TargetRegisterClass *RC,
- SmallVectorImpl<MachineInstr *> &NewMIs) const;
- void LoadRegFromStackSlot(MachineFunction &MF, const DebugLoc &DL,
- unsigned DestReg, int FrameIdx,
- const TargetRegisterClass *RC,
- SmallVectorImpl<MachineInstr *> &NewMIs) const;
- // Replace the instruction with single LI if possible. \p DefMI must be LI or
- // LI8.
- bool simplifyToLI(MachineInstr &MI, MachineInstr &DefMI,
- unsigned OpNoForForwarding, MachineInstr **KilledDef) const;
- // If the inst is imm-form and its register operand is produced by a ADDI, put
- // the imm into the inst directly and remove the ADDI if possible.
- bool transformToNewImmFormFedByAdd(MachineInstr &MI, MachineInstr &DefMI,
- unsigned OpNoForForwarding) const;
- // If the inst is x-form and has imm-form and one of its operand is produced
- // by a LI, put the imm into the inst directly and remove the LI if possible.
- bool transformToImmFormFedByLI(MachineInstr &MI, const ImmInstrInfo &III,
- unsigned ConstantOpNo,
- MachineInstr &DefMI) const;
- // If the inst is x-form and has imm-form and one of its operand is produced
- // by an add-immediate, try to transform it when possible.
- bool transformToImmFormFedByAdd(MachineInstr &MI, const ImmInstrInfo &III,
- unsigned ConstantOpNo, MachineInstr &DefMI,
- bool KillDefMI) const;
- // Try to find that, if the instruction 'MI' contains any operand that
- // could be forwarded from some inst that feeds it. If yes, return the
- // Def of that operand. And OpNoForForwarding is the operand index in
- // the 'MI' for that 'Def'. If we see another use of this Def between
- // the Def and the MI, SeenIntermediateUse becomes 'true'.
- MachineInstr *getForwardingDefMI(MachineInstr &MI,
- unsigned &OpNoForForwarding,
- bool &SeenIntermediateUse) const;
- // Can the user MI have it's source at index \p OpNoForForwarding
- // forwarded from an add-immediate that feeds it?
- bool isUseMIElgibleForForwarding(MachineInstr &MI, const ImmInstrInfo &III,
- unsigned OpNoForForwarding) const;
- bool isDefMIElgibleForForwarding(MachineInstr &DefMI,
- const ImmInstrInfo &III,
- MachineOperand *&ImmMO,
- MachineOperand *&RegMO) const;
- bool isImmElgibleForForwarding(const MachineOperand &ImmMO,
- const MachineInstr &DefMI,
- const ImmInstrInfo &III,
- int64_t &Imm,
- int64_t BaseImm = 0) const;
- bool isRegElgibleForForwarding(const MachineOperand &RegMO,
- const MachineInstr &DefMI,
- const MachineInstr &MI, bool KillDefMI,
- bool &IsFwdFeederRegKilled) const;
- unsigned getSpillTarget() const;
- const unsigned *getStoreOpcodesForSpillArray() const;
- const unsigned *getLoadOpcodesForSpillArray() const;
- unsigned getSpillIndex(const TargetRegisterClass *RC) const;
- int16_t getFMAOpIdxInfo(unsigned Opcode) const;
- void reassociateFMA(MachineInstr &Root, MachineCombinerPattern Pattern,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs,
- DenseMap<unsigned, unsigned> &InstrIdxForVirtReg) const;
- bool isLoadFromConstantPool(MachineInstr *I) const;
- Register
- generateLoadForNewConst(unsigned Idx, MachineInstr *MI, Type *Ty,
- SmallVectorImpl<MachineInstr *> &InsInstrs) const;
- const Constant *getConstantFromConstantPool(MachineInstr *I) const;
- virtual void anchor();
- protected:
- /// Commutes the operands in the given instruction.
- /// The commutable operands are specified by their indices OpIdx1 and OpIdx2.
- ///
- /// Do not call this method for a non-commutable instruction or for
- /// non-commutable pair of operand indices OpIdx1 and OpIdx2.
- /// Even though the instruction is commutable, the method may still
- /// fail to commute the operands, null pointer is returned in such cases.
- ///
- /// For example, we can commute rlwimi instructions, but only if the
- /// rotate amt is zero. We also have to munge the immediates a bit.
- MachineInstr *commuteInstructionImpl(MachineInstr &MI, bool NewMI,
- unsigned OpIdx1,
- unsigned OpIdx2) const override;
- public:
- explicit PPCInstrInfo(PPCSubtarget &STI);
- /// getRegisterInfo - TargetInstrInfo is a superset of MRegister info. As
- /// such, whenever a client has an instance of instruction info, it should
- /// always be able to get register info as well (through this method).
- ///
- const PPCRegisterInfo &getRegisterInfo() const { return RI; }
- bool isXFormMemOp(unsigned Opcode) const {
- return get(Opcode).TSFlags & PPCII::XFormMemOp;
- }
- bool isPrefixed(unsigned Opcode) const {
- return get(Opcode).TSFlags & PPCII::Prefixed;
- }
- static bool isSameClassPhysRegCopy(unsigned Opcode) {
- unsigned CopyOpcodes[] = {PPC::OR, PPC::OR8, PPC::FMR,
- PPC::VOR, PPC::XXLOR, PPC::XXLORf,
- PPC::XSCPSGNDP, PPC::MCRF, PPC::CROR,
- PPC::EVOR, -1U};
- for (int i = 0; CopyOpcodes[i] != -1U; i++)
- if (Opcode == CopyOpcodes[i])
- return true;
- return false;
- }
- ScheduleHazardRecognizer *
- CreateTargetHazardRecognizer(const TargetSubtargetInfo *STI,
- const ScheduleDAG *DAG) const override;
- ScheduleHazardRecognizer *
- CreateTargetPostRAHazardRecognizer(const InstrItineraryData *II,
- const ScheduleDAG *DAG) const override;
- unsigned getInstrLatency(const InstrItineraryData *ItinData,
- const MachineInstr &MI,
- unsigned *PredCost = nullptr) const override;
- int getOperandLatency(const InstrItineraryData *ItinData,
- const MachineInstr &DefMI, unsigned DefIdx,
- const MachineInstr &UseMI,
- unsigned UseIdx) const override;
- int getOperandLatency(const InstrItineraryData *ItinData,
- SDNode *DefNode, unsigned DefIdx,
- SDNode *UseNode, unsigned UseIdx) const override {
- return PPCGenInstrInfo::getOperandLatency(ItinData, DefNode, DefIdx,
- UseNode, UseIdx);
- }
- bool hasLowDefLatency(const TargetSchedModel &SchedModel,
- const MachineInstr &DefMI,
- unsigned DefIdx) const override {
- // Machine LICM should hoist all instructions in low-register-pressure
- // situations; none are sufficiently free to justify leaving in a loop
- // body.
- return false;
- }
- bool useMachineCombiner() const override {
- return true;
- }
- /// When getMachineCombinerPatterns() finds patterns, this function generates
- /// the instructions that could replace the original code sequence
- void genAlternativeCodeSequence(
- MachineInstr &Root, MachineCombinerPattern Pattern,
- SmallVectorImpl<MachineInstr *> &InsInstrs,
- SmallVectorImpl<MachineInstr *> &DelInstrs,
- DenseMap<unsigned, unsigned> &InstrIdxForVirtReg) const override;
- /// Return true when there is potentially a faster code sequence for a fma
- /// chain ending in \p Root. All potential patterns are output in the \p
- /// P array.
- bool getFMAPatterns(MachineInstr &Root,
- SmallVectorImpl<MachineCombinerPattern> &P,
- bool DoRegPressureReduce) const;
- /// Return true when there is potentially a faster code sequence
- /// for an instruction chain ending in <Root>. All potential patterns are
- /// output in the <Pattern> array.
- bool getMachineCombinerPatterns(MachineInstr &Root,
- SmallVectorImpl<MachineCombinerPattern> &P,
- bool DoRegPressureReduce) const override;
- /// On PowerPC, we leverage machine combiner pass to reduce register pressure
- /// when the register pressure is high for one BB.
- /// Return true if register pressure for \p MBB is high and ABI is supported
- /// to reduce register pressure. Otherwise return false.
- bool
- shouldReduceRegisterPressure(MachineBasicBlock *MBB,
- RegisterClassInfo *RegClassInfo) const override;
- /// Fixup the placeholders we put in genAlternativeCodeSequence() for
- /// MachineCombiner.
- void
- finalizeInsInstrs(MachineInstr &Root, MachineCombinerPattern &P,
- SmallVectorImpl<MachineInstr *> &InsInstrs) const override;
- bool isAssociativeAndCommutative(const MachineInstr &Inst) const override;
- /// On PowerPC, we try to reassociate FMA chain which will increase
- /// instruction size. Set extension resource length limit to 1 for edge case.
- /// Resource Length is calculated by scaled resource usage in getCycles().
- /// Because of the division in getCycles(), it returns different cycles due to
- /// legacy scaled resource usage. So new resource length may be same with
- /// legacy or 1 bigger than legacy.
- /// We need to execlude the 1 bigger case even the resource length is not
- /// perserved for more FMA chain reassociations on PowerPC.
- int getExtendResourceLenLimit() const override { return 1; }
- void setSpecialOperandAttr(MachineInstr &OldMI1, MachineInstr &OldMI2,
- MachineInstr &NewMI1,
- MachineInstr &NewMI2) const override;
- // PowerPC specific version of setSpecialOperandAttr that copies Flags to MI
- // and clears nuw, nsw, and exact flags.
- void setSpecialOperandAttr(MachineInstr &MI, uint16_t Flags) const;
- bool isCoalescableExtInstr(const MachineInstr &MI,
- Register &SrcReg, Register &DstReg,
- unsigned &SubIdx) const override;
- unsigned isLoadFromStackSlot(const MachineInstr &MI,
- int &FrameIndex) const override;
- bool isReallyTriviallyReMaterializable(const MachineInstr &MI,
- AAResults *AA) const override;
- unsigned isStoreToStackSlot(const MachineInstr &MI,
- int &FrameIndex) const override;
- bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1,
- unsigned &SrcOpIdx2) const override;
- void insertNoop(MachineBasicBlock &MBB,
- MachineBasicBlock::iterator MI) const override;
- // Branch analysis.
- bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB,
- MachineBasicBlock *&FBB,
- SmallVectorImpl<MachineOperand> &Cond,
- bool AllowModify) const override;
- unsigned removeBranch(MachineBasicBlock &MBB,
- int *BytesRemoved = nullptr) const override;
- unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB,
- MachineBasicBlock *FBB, ArrayRef<MachineOperand> Cond,
- const DebugLoc &DL,
- int *BytesAdded = nullptr) const override;
- // Select analysis.
- bool canInsertSelect(const MachineBasicBlock &, ArrayRef<MachineOperand> Cond,
- Register, Register, Register, int &, int &,
- int &) const override;
- void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI,
- const DebugLoc &DL, Register DstReg,
- ArrayRef<MachineOperand> Cond, Register TrueReg,
- Register FalseReg) const override;
- void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
- const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg,
- bool KillSrc) const override;
- void storeRegToStackSlot(MachineBasicBlock &MBB,
- MachineBasicBlock::iterator MBBI,
- Register SrcReg, bool isKill, int FrameIndex,
- const TargetRegisterClass *RC,
- const TargetRegisterInfo *TRI) const override;
- // Emits a register spill without updating the register class for vector
- // registers. This ensures that when we spill a vector register the
- // element order in the register is the same as it was in memory.
- void storeRegToStackSlotNoUpd(MachineBasicBlock &MBB,
- MachineBasicBlock::iterator MBBI,
- unsigned SrcReg, bool isKill, int FrameIndex,
- const TargetRegisterClass *RC,
- const TargetRegisterInfo *TRI) const;
- void loadRegFromStackSlot(MachineBasicBlock &MBB,
- MachineBasicBlock::iterator MBBI,
- Register DestReg, int FrameIndex,
- const TargetRegisterClass *RC,
- const TargetRegisterInfo *TRI) const override;
- // Emits a register reload without updating the register class for vector
- // registers. This ensures that when we reload a vector register the
- // element order in the register is the same as it was in memory.
- void loadRegFromStackSlotNoUpd(MachineBasicBlock &MBB,
- MachineBasicBlock::iterator MBBI,
- unsigned DestReg, int FrameIndex,
- const TargetRegisterClass *RC,
- const TargetRegisterInfo *TRI) const;
- unsigned getStoreOpcodeForSpill(const TargetRegisterClass *RC) const;
- unsigned getLoadOpcodeForSpill(const TargetRegisterClass *RC) const;
- bool
- reverseBranchCondition(SmallVectorImpl<MachineOperand> &Cond) const override;
- bool FoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI, Register Reg,
- MachineRegisterInfo *MRI) const override;
- bool onlyFoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI,
- Register Reg) const;
- // If conversion by predication (only supported by some branch instructions).
- // All of the profitability checks always return true; it is always
- // profitable to use the predicated branches.
- bool isProfitableToIfCvt(MachineBasicBlock &MBB,
- unsigned NumCycles, unsigned ExtraPredCycles,
- BranchProbability Probability) const override {
- return true;
- }
- bool isProfitableToIfCvt(MachineBasicBlock &TMBB,
- unsigned NumT, unsigned ExtraT,
- MachineBasicBlock &FMBB,
- unsigned NumF, unsigned ExtraF,
- BranchProbability Probability) const override;
- bool isProfitableToDupForIfCvt(MachineBasicBlock &MBB, unsigned NumCycles,
- BranchProbability Probability) const override {
- return true;
- }
- bool isProfitableToUnpredicate(MachineBasicBlock &TMBB,
- MachineBasicBlock &FMBB) const override {
- return false;
- }
- // Predication support.
- bool isPredicated(const MachineInstr &MI) const override;
- bool isSchedulingBoundary(const MachineInstr &MI,
- const MachineBasicBlock *MBB,
- const MachineFunction &MF) const override;
- bool PredicateInstruction(MachineInstr &MI,
- ArrayRef<MachineOperand> Pred) const override;
- bool SubsumesPredicate(ArrayRef<MachineOperand> Pred1,
- ArrayRef<MachineOperand> Pred2) const override;
- bool ClobbersPredicate(MachineInstr &MI, std::vector<MachineOperand> &Pred,
- bool SkipDead) const override;
- // Comparison optimization.
- bool analyzeCompare(const MachineInstr &MI, Register &SrcReg,
- Register &SrcReg2, int64_t &Mask,
- int64_t &Value) const override;
- bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg,
- Register SrcReg2, int64_t Mask, int64_t Value,
- const MachineRegisterInfo *MRI) const override;
- /// Return true if get the base operand, byte offset of an instruction and
- /// the memory width. Width is the size of memory that is being
- /// loaded/stored (e.g. 1, 2, 4, 8).
- bool getMemOperandWithOffsetWidth(const MachineInstr &LdSt,
- const MachineOperand *&BaseOp,
- int64_t &Offset, unsigned &Width,
- const TargetRegisterInfo *TRI) const;
- /// Get the base operand and byte offset of an instruction that reads/writes
- /// memory.
- bool getMemOperandsWithOffsetWidth(
- const MachineInstr &LdSt,
- SmallVectorImpl<const MachineOperand *> &BaseOps, int64_t &Offset,
- bool &OffsetIsScalable, unsigned &Width,
- const TargetRegisterInfo *TRI) const override;
- /// Returns true if the two given memory operations should be scheduled
- /// adjacent.
- bool shouldClusterMemOps(ArrayRef<const MachineOperand *> BaseOps1,
- ArrayRef<const MachineOperand *> BaseOps2,
- unsigned NumLoads, unsigned NumBytes) const override;
- /// Return true if two MIs access different memory addresses and false
- /// otherwise
- bool
- areMemAccessesTriviallyDisjoint(const MachineInstr &MIa,
- const MachineInstr &MIb) const override;
- /// GetInstSize - Return the number of bytes of code the specified
- /// instruction may be. This returns the maximum number of bytes.
- ///
- unsigned getInstSizeInBytes(const MachineInstr &MI) const override;
- MCInst getNop() const override;
- std::pair<unsigned, unsigned>
- decomposeMachineOperandsTargetFlags(unsigned TF) const override;
- ArrayRef<std::pair<unsigned, const char *>>
- getSerializableDirectMachineOperandTargetFlags() const override;
- ArrayRef<std::pair<unsigned, const char *>>
- getSerializableBitmaskMachineOperandTargetFlags() const override;
- // Expand VSX Memory Pseudo instruction to either a VSX or a FP instruction.
- bool expandVSXMemPseudo(MachineInstr &MI) const;
- // Lower pseudo instructions after register allocation.
- bool expandPostRAPseudo(MachineInstr &MI) const override;
- static bool isVFRegister(unsigned Reg) {
- return Reg >= PPC::VF0 && Reg <= PPC::VF31;
- }
- static bool isVRRegister(unsigned Reg) {
- return Reg >= PPC::V0 && Reg <= PPC::V31;
- }
- const TargetRegisterClass *updatedRC(const TargetRegisterClass *RC) const;
- static int getRecordFormOpcode(unsigned Opcode);
- bool isTOCSaveMI(const MachineInstr &MI) const;
- bool isSignOrZeroExtended(const MachineInstr &MI, bool SignExt,
- const unsigned PhiDepth) const;
- /// Return true if the output of the instruction is always a sign-extended,
- /// i.e. 0 to 31-th bits are same as 32-th bit.
- bool isSignExtended(const MachineInstr &MI, const unsigned depth = 0) const {
- return isSignOrZeroExtended(MI, true, depth);
- }
- /// Return true if the output of the instruction is always zero-extended,
- /// i.e. 0 to 31-th bits are all zeros
- bool isZeroExtended(const MachineInstr &MI, const unsigned depth = 0) const {
- return isSignOrZeroExtended(MI, false, depth);
- }
- bool convertToImmediateForm(MachineInstr &MI,
- MachineInstr **KilledDef = nullptr) const;
- bool foldFrameOffset(MachineInstr &MI) const;
- bool combineRLWINM(MachineInstr &MI, MachineInstr **ToErase = nullptr) const;
- bool isADDIInstrEligibleForFolding(MachineInstr &ADDIMI, int64_t &Imm) const;
- bool isADDInstrEligibleForFolding(MachineInstr &ADDMI) const;
- bool isImmInstrEligibleForFolding(MachineInstr &MI, unsigned &BaseReg,
- unsigned &XFormOpcode,
- int64_t &OffsetOfImmInstr,
- ImmInstrInfo &III) const;
- bool isValidToBeChangedReg(MachineInstr *ADDMI, unsigned Index,
- MachineInstr *&ADDIMI, int64_t &OffsetAddi,
- int64_t OffsetImm) const;
- /// Fixup killed/dead flag for register \p RegNo between instructions [\p
- /// StartMI, \p EndMI]. Some pre-RA or post-RA transformations may violate
- /// register killed/dead flags semantics, this function can be called to fix
- /// up. Before calling this function,
- /// 1. Ensure that \p RegNo liveness is killed after instruction \p EndMI.
- /// 2. Ensure that there is no new definition between (\p StartMI, \p EndMI)
- /// and possible definition for \p RegNo is \p StartMI or \p EndMI. For
- /// pre-RA cases, definition may be \p StartMI through COPY, \p StartMI
- /// will be adjust to true definition.
- /// 3. We can do accurate fixup for the case when all instructions between
- /// [\p StartMI, \p EndMI] are in same basic block.
- /// 4. For the case when \p StartMI and \p EndMI are not in same basic block,
- /// we conservatively clear kill flag for all uses of \p RegNo for pre-RA
- /// and for post-RA, we give an assertion as without reaching definition
- /// analysis post-RA, \p StartMI and \p EndMI are hard to keep right.
- void fixupIsDeadOrKill(MachineInstr *StartMI, MachineInstr *EndMI,
- unsigned RegNo) const;
- void replaceInstrWithLI(MachineInstr &MI, const LoadImmediateInfo &LII) const;
- void replaceInstrOperandWithImm(MachineInstr &MI, unsigned OpNo,
- int64_t Imm) const;
- bool instrHasImmForm(unsigned Opc, bool IsVFReg, ImmInstrInfo &III,
- bool PostRA) const;
- // In PostRA phase, try to find instruction defines \p Reg before \p MI.
- // \p SeenIntermediate is set to true if uses between DefMI and \p MI exist.
- MachineInstr *getDefMIPostRA(unsigned Reg, MachineInstr &MI,
- bool &SeenIntermediateUse) const;
- /// getRegNumForOperand - some operands use different numbering schemes
- /// for the same registers. For example, a VSX instruction may have any of
- /// vs0-vs63 allocated whereas an Altivec instruction could only have
- /// vs32-vs63 allocated (numbered as v0-v31). This function returns the actual
- /// register number needed for the opcode/operand number combination.
- /// The operand number argument will be useful when we need to extend this
- /// to instructions that use both Altivec and VSX numbering (for different
- /// operands).
- static unsigned getRegNumForOperand(const MCInstrDesc &Desc, unsigned Reg,
- unsigned OpNo) {
- int16_t regClass = Desc.OpInfo[OpNo].RegClass;
- switch (regClass) {
- // We store F0-F31, VF0-VF31 in MCOperand and it should be F0-F31,
- // VSX32-VSX63 during encoding/disassembling
- case PPC::VSSRCRegClassID:
- case PPC::VSFRCRegClassID:
- if (isVFRegister(Reg))
- return PPC::VSX32 + (Reg - PPC::VF0);
- break;
- // We store VSL0-VSL31, V0-V31 in MCOperand and it should be VSL0-VSL31,
- // VSX32-VSX63 during encoding/disassembling
- case PPC::VSRCRegClassID:
- if (isVRRegister(Reg))
- return PPC::VSX32 + (Reg - PPC::V0);
- break;
- // Other RegClass doesn't need mapping
- default:
- break;
- }
- return Reg;
- }
- /// Check \p Opcode is BDNZ (Decrement CTR and branch if it is still nonzero).
- bool isBDNZ(unsigned Opcode) const;
- /// Find the hardware loop instruction used to set-up the specified loop.
- /// On PPC, we have two instructions used to set-up the hardware loop
- /// (MTCTRloop, MTCTR8loop) with corresponding endloop (BDNZ, BDNZ8)
- /// instructions to indicate the end of a loop.
- MachineInstr *
- findLoopInstr(MachineBasicBlock &PreHeader,
- SmallPtrSet<MachineBasicBlock *, 8> &Visited) const;
- /// Analyze loop L, which must be a single-basic-block loop, and if the
- /// conditions can be understood enough produce a PipelinerLoopInfo object.
- std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
- analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override;
- };
- }
- #endif
|