123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251 |
- //===- PPCMachineScheduler.cpp - MI Scheduler for PowerPC -------------===//
- //
- // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
- // See https://llvm.org/LICENSE.txt for license information.
- // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
- //
- //===----------------------------------------------------------------------===//
- #include "PPCMachineScheduler.h"
- #include "MCTargetDesc/PPCMCTargetDesc.h"
- using namespace llvm;
- static cl::opt<bool>
- DisableAddiLoadHeuristic("disable-ppc-sched-addi-load",
- cl::desc("Disable scheduling addi instruction before"
- "load for ppc"), cl::Hidden);
- static cl::opt<bool>
- EnableAddiHeuristic("ppc-postra-bias-addi",
- cl::desc("Enable scheduling addi instruction as early"
- "as possible post ra"),
- cl::Hidden, cl::init(true));
- static bool isADDIInstr(const GenericScheduler::SchedCandidate &Cand) {
- return Cand.SU->getInstr()->getOpcode() == PPC::ADDI ||
- Cand.SU->getInstr()->getOpcode() == PPC::ADDI8;
- }
- bool PPCPreRASchedStrategy::biasAddiLoadCandidate(SchedCandidate &Cand,
- SchedCandidate &TryCand,
- SchedBoundary &Zone) const {
- if (DisableAddiLoadHeuristic)
- return false;
- SchedCandidate &FirstCand = Zone.isTop() ? TryCand : Cand;
- SchedCandidate &SecondCand = Zone.isTop() ? Cand : TryCand;
- if (isADDIInstr(FirstCand) && SecondCand.SU->getInstr()->mayLoad()) {
- TryCand.Reason = Stall;
- return true;
- }
- if (FirstCand.SU->getInstr()->mayLoad() && isADDIInstr(SecondCand)) {
- TryCand.Reason = NoCand;
- return true;
- }
- return false;
- }
- bool PPCPreRASchedStrategy::tryCandidate(SchedCandidate &Cand,
- SchedCandidate &TryCand,
- SchedBoundary *Zone) const {
- // From GenericScheduler::tryCandidate
- // Initialize the candidate if needed.
- if (!Cand.isValid()) {
- TryCand.Reason = NodeOrder;
- return true;
- }
- // Bias PhysReg Defs and copies to their uses and defined respectively.
- if (tryGreater(biasPhysReg(TryCand.SU, TryCand.AtTop),
- biasPhysReg(Cand.SU, Cand.AtTop), TryCand, Cand, PhysReg))
- return TryCand.Reason != NoCand;
- // Avoid exceeding the target's limit.
- if (DAG->isTrackingPressure() &&
- tryPressure(TryCand.RPDelta.Excess, Cand.RPDelta.Excess, TryCand, Cand,
- RegExcess, TRI, DAG->MF))
- return TryCand.Reason != NoCand;
- // Avoid increasing the max critical pressure in the scheduled region.
- if (DAG->isTrackingPressure() &&
- tryPressure(TryCand.RPDelta.CriticalMax, Cand.RPDelta.CriticalMax,
- TryCand, Cand, RegCritical, TRI, DAG->MF))
- return TryCand.Reason != NoCand;
- // We only compare a subset of features when comparing nodes between
- // Top and Bottom boundary. Some properties are simply incomparable, in many
- // other instances we should only override the other boundary if something
- // is a clear good pick on one boundary. Skip heuristics that are more
- // "tie-breaking" in nature.
- bool SameBoundary = Zone != nullptr;
- if (SameBoundary) {
- // For loops that are acyclic path limited, aggressively schedule for
- // latency. Within an single cycle, whenever CurrMOps > 0, allow normal
- // heuristics to take precedence.
- if (Rem.IsAcyclicLatencyLimited && !Zone->getCurrMOps() &&
- tryLatency(TryCand, Cand, *Zone))
- return TryCand.Reason != NoCand;
- // Prioritize instructions that read unbuffered resources by stall cycles.
- if (tryLess(Zone->getLatencyStallCycles(TryCand.SU),
- Zone->getLatencyStallCycles(Cand.SU), TryCand, Cand, Stall))
- return TryCand.Reason != NoCand;
- }
- // Keep clustered nodes together to encourage downstream peephole
- // optimizations which may reduce resource requirements.
- //
- // This is a best effort to set things up for a post-RA pass. Optimizations
- // like generating loads of multiple registers should ideally be done within
- // the scheduler pass by combining the loads during DAG postprocessing.
- const SUnit *CandNextClusterSU =
- Cand.AtTop ? DAG->getNextClusterSucc() : DAG->getNextClusterPred();
- const SUnit *TryCandNextClusterSU =
- TryCand.AtTop ? DAG->getNextClusterSucc() : DAG->getNextClusterPred();
- if (tryGreater(TryCand.SU == TryCandNextClusterSU,
- Cand.SU == CandNextClusterSU, TryCand, Cand, Cluster))
- return TryCand.Reason != NoCand;
- if (SameBoundary) {
- // Weak edges are for clustering and other constraints.
- if (tryLess(getWeakLeft(TryCand.SU, TryCand.AtTop),
- getWeakLeft(Cand.SU, Cand.AtTop), TryCand, Cand, Weak))
- return TryCand.Reason != NoCand;
- }
- // Avoid increasing the max pressure of the entire region.
- if (DAG->isTrackingPressure() &&
- tryPressure(TryCand.RPDelta.CurrentMax, Cand.RPDelta.CurrentMax, TryCand,
- Cand, RegMax, TRI, DAG->MF))
- return TryCand.Reason != NoCand;
- if (SameBoundary) {
- // Avoid critical resource consumption and balance the schedule.
- TryCand.initResourceDelta(DAG, SchedModel);
- if (tryLess(TryCand.ResDelta.CritResources, Cand.ResDelta.CritResources,
- TryCand, Cand, ResourceReduce))
- return TryCand.Reason != NoCand;
- if (tryGreater(TryCand.ResDelta.DemandedResources,
- Cand.ResDelta.DemandedResources, TryCand, Cand,
- ResourceDemand))
- return TryCand.Reason != NoCand;
- // Avoid serializing long latency dependence chains.
- // For acyclic path limited loops, latency was already checked above.
- if (!RegionPolicy.DisableLatencyHeuristic && TryCand.Policy.ReduceLatency &&
- !Rem.IsAcyclicLatencyLimited && tryLatency(TryCand, Cand, *Zone))
- return TryCand.Reason != NoCand;
- // Fall through to original instruction order.
- if ((Zone->isTop() && TryCand.SU->NodeNum < Cand.SU->NodeNum) ||
- (!Zone->isTop() && TryCand.SU->NodeNum > Cand.SU->NodeNum)) {
- TryCand.Reason = NodeOrder;
- }
- }
- // GenericScheduler::tryCandidate end
- // Add powerpc specific heuristic only when TryCand isn't selected or
- // selected as node order.
- if (TryCand.Reason != NodeOrder && TryCand.Reason != NoCand)
- return true;
- // There are some benefits to schedule the ADDI before the load to hide the
- // latency, as RA may create a true dependency between the load and addi.
- if (SameBoundary) {
- if (biasAddiLoadCandidate(Cand, TryCand, *Zone))
- return TryCand.Reason != NoCand;
- }
- return TryCand.Reason != NoCand;
- }
- bool PPCPostRASchedStrategy::biasAddiCandidate(SchedCandidate &Cand,
- SchedCandidate &TryCand) const {
- if (!EnableAddiHeuristic)
- return false;
- if (isADDIInstr(TryCand) && !isADDIInstr(Cand)) {
- TryCand.Reason = Stall;
- return true;
- }
- return false;
- }
- bool PPCPostRASchedStrategy::tryCandidate(SchedCandidate &Cand,
- SchedCandidate &TryCand) {
- // From PostGenericScheduler::tryCandidate
- // Initialize the candidate if needed.
- if (!Cand.isValid()) {
- TryCand.Reason = NodeOrder;
- return true;
- }
- // Prioritize instructions that read unbuffered resources by stall cycles.
- if (tryLess(Top.getLatencyStallCycles(TryCand.SU),
- Top.getLatencyStallCycles(Cand.SU), TryCand, Cand, Stall))
- return TryCand.Reason != NoCand;
- // Keep clustered nodes together.
- if (tryGreater(TryCand.SU == DAG->getNextClusterSucc(),
- Cand.SU == DAG->getNextClusterSucc(), TryCand, Cand, Cluster))
- return TryCand.Reason != NoCand;
- // Avoid critical resource consumption and balance the schedule.
- if (tryLess(TryCand.ResDelta.CritResources, Cand.ResDelta.CritResources,
- TryCand, Cand, ResourceReduce))
- return TryCand.Reason != NoCand;
- if (tryGreater(TryCand.ResDelta.DemandedResources,
- Cand.ResDelta.DemandedResources, TryCand, Cand,
- ResourceDemand))
- return TryCand.Reason != NoCand;
- // Avoid serializing long latency dependence chains.
- if (Cand.Policy.ReduceLatency && tryLatency(TryCand, Cand, Top)) {
- return TryCand.Reason != NoCand;
- }
- // Fall through to original instruction order.
- if (TryCand.SU->NodeNum < Cand.SU->NodeNum)
- TryCand.Reason = NodeOrder;
- // PostGenericScheduler::tryCandidate end
- // Add powerpc post ra specific heuristic only when TryCand isn't selected or
- // selected as node order.
- if (TryCand.Reason != NodeOrder && TryCand.Reason != NoCand)
- return true;
- // There are some benefits to schedule the ADDI as early as possible post ra
- // to avoid stalled by vector instructions which take up all the hw units.
- // And ADDI is usually used to post inc the loop indvar, which matters the
- // performance.
- if (biasAddiCandidate(Cand, TryCand))
- return TryCand.Reason != NoCand;
- return TryCand.Reason != NoCand;
- }
- void PPCPostRASchedStrategy::enterMBB(MachineBasicBlock *MBB) {
- // Custom PPC PostRA specific behavior here.
- PostGenericScheduler::enterMBB(MBB);
- }
- void PPCPostRASchedStrategy::leaveMBB() {
- // Custom PPC PostRA specific behavior here.
- PostGenericScheduler::leaveMBB();
- }
- void PPCPostRASchedStrategy::initialize(ScheduleDAGMI *Dag) {
- // Custom PPC PostRA specific initialization here.
- PostGenericScheduler::initialize(Dag);
- }
- SUnit *PPCPostRASchedStrategy::pickNode(bool &IsTopNode) {
- // Custom PPC PostRA specific scheduling here.
- return PostGenericScheduler::pickNode(IsTopNode);
- }
|