1//===--- AArch64StorePairSuppress.cpp --- Suppress store pair formation ---===// 2// 3// The LLVM Compiler Infrastructure 4// 5// This file is distributed under the University of Illinois Open Source 6// License. See LICENSE.TXT for details. 7// 8//===----------------------------------------------------------------------===// 9// 10// This pass identifies floating point stores that should not be combined into 11// store pairs. Later we may do the same for floating point loads. 12// ===---------------------------------------------------------------------===// 13 14#include "AArch64InstrInfo.h" 15#include "llvm/CodeGen/MachineFunction.h" 16#include "llvm/CodeGen/MachineFunctionPass.h" 17#include "llvm/CodeGen/MachineInstr.h" 18#include "llvm/CodeGen/MachineTraceMetrics.h" 19#include "llvm/CodeGen/TargetSchedule.h" 20#include "llvm/Support/Debug.h" 21#include "llvm/Support/raw_ostream.h" 22#include "llvm/Target/TargetInstrInfo.h" 23 24using namespace llvm; 25 26#define DEBUG_TYPE "aarch64-stp-suppress" 27 28namespace { 29class AArch64StorePairSuppress : public MachineFunctionPass { 30 const AArch64InstrInfo *TII; 31 const TargetRegisterInfo *TRI; 32 const MachineRegisterInfo *MRI; 33 TargetSchedModel SchedModel; 34 MachineTraceMetrics *Traces; 35 MachineTraceMetrics::Ensemble *MinInstr; 36 37public: 38 static char ID; 39 AArch64StorePairSuppress() : MachineFunctionPass(ID) {} 40 41 const char *getPassName() const override { 42 return "AArch64 Store Pair Suppression"; 43 } 44 45 bool runOnMachineFunction(MachineFunction &F) override; 46 47private: 48 bool shouldAddSTPToBlock(const MachineBasicBlock *BB); 49 50 bool isNarrowFPStore(const MachineInstr &MI); 51 52 void getAnalysisUsage(AnalysisUsage &AU) const override { 53 AU.setPreservesCFG(); 54 AU.addRequired<MachineTraceMetrics>(); 55 AU.addPreserved<MachineTraceMetrics>(); 56 MachineFunctionPass::getAnalysisUsage(AU); 57 } 58}; 59char AArch64StorePairSuppress::ID = 0; 60} // anonymous 61 62FunctionPass *llvm::createAArch64StorePairSuppressPass() { 63 return new AArch64StorePairSuppress(); 64} 65 66/// Return true if an STP can be added to this block without increasing the 67/// critical resource height. STP is good to form in Ld/St limited blocks and 68/// bad to form in float-point limited blocks. This is true independent of the 69/// critical path. If the critical path is longer than the resource height, the 70/// extra vector ops can limit physreg renaming. Otherwise, it could simply 71/// oversaturate the vector units. 72bool AArch64StorePairSuppress::shouldAddSTPToBlock(const MachineBasicBlock *BB) { 73 if (!MinInstr) 74 MinInstr = Traces->getEnsemble(MachineTraceMetrics::TS_MinInstrCount); 75 76 MachineTraceMetrics::Trace BBTrace = MinInstr->getTrace(BB); 77 unsigned ResLength = BBTrace.getResourceLength(); 78 79 // Get the machine model's scheduling class for STPQi. 80 // Bypass TargetSchedule's SchedClass resolution since we only have an opcode. 81 unsigned SCIdx = TII->get(AArch64::STPDi).getSchedClass(); 82 const MCSchedClassDesc *SCDesc = 83 SchedModel.getMCSchedModel()->getSchedClassDesc(SCIdx); 84 85 // If a subtarget does not define resources for STPQi, bail here. 86 if (SCDesc->isValid() && !SCDesc->isVariant()) { 87 unsigned ResLenWithSTP = BBTrace.getResourceLength(None, SCDesc); 88 if (ResLenWithSTP > ResLength) { 89 DEBUG(dbgs() << " Suppress STP in BB: " << BB->getNumber() 90 << " resources " << ResLength << " -> " << ResLenWithSTP 91 << "\n"); 92 return false; 93 } 94 } 95 return true; 96} 97 98/// Return true if this is a floating-point store smaller than the V reg. On 99/// cyclone, these require a vector shuffle before storing a pair. 100/// Ideally we would call getMatchingPairOpcode() and have the machine model 101/// tell us if it's profitable with no cpu knowledge here. 102/// 103/// FIXME: We plan to develop a decent Target abstraction for simple loads and 104/// stores. Until then use a nasty switch similar to AArch64LoadStoreOptimizer. 105bool AArch64StorePairSuppress::isNarrowFPStore(const MachineInstr &MI) { 106 switch (MI.getOpcode()) { 107 default: 108 return false; 109 case AArch64::STRSui: 110 case AArch64::STRDui: 111 case AArch64::STURSi: 112 case AArch64::STURDi: 113 return true; 114 } 115} 116 117bool AArch64StorePairSuppress::runOnMachineFunction(MachineFunction &MF) { 118 if (skipFunction(*MF.getFunction())) 119 return false; 120 121 const TargetSubtargetInfo &ST = MF.getSubtarget(); 122 TII = static_cast<const AArch64InstrInfo *>(ST.getInstrInfo()); 123 TRI = ST.getRegisterInfo(); 124 MRI = &MF.getRegInfo(); 125 SchedModel.init(ST.getSchedModel(), &ST, TII); 126 Traces = &getAnalysis<MachineTraceMetrics>(); 127 MinInstr = nullptr; 128 129 DEBUG(dbgs() << "*** " << getPassName() << ": " << MF.getName() << '\n'); 130 131 if (!SchedModel.hasInstrSchedModel()) { 132 DEBUG(dbgs() << " Skipping pass: no machine model present.\n"); 133 return false; 134 } 135 136 // Check for a sequence of stores to the same base address. We don't need to 137 // precisely determine whether a store pair can be formed. But we do want to 138 // filter out most situations where we can't form store pairs to avoid 139 // computing trace metrics in those cases. 140 for (auto &MBB : MF) { 141 bool SuppressSTP = false; 142 unsigned PrevBaseReg = 0; 143 for (auto &MI : MBB) { 144 if (!isNarrowFPStore(MI)) 145 continue; 146 unsigned BaseReg; 147 int64_t Offset; 148 if (TII->getMemOpBaseRegImmOfs(MI, BaseReg, Offset, TRI)) { 149 if (PrevBaseReg == BaseReg) { 150 // If this block can take STPs, skip ahead to the next block. 151 if (!SuppressSTP && shouldAddSTPToBlock(MI.getParent())) 152 break; 153 // Otherwise, continue unpairing the stores in this block. 154 DEBUG(dbgs() << "Unpairing store " << MI << "\n"); 155 SuppressSTP = true; 156 TII->suppressLdStPair(MI); 157 } 158 PrevBaseReg = BaseReg; 159 } else 160 PrevBaseReg = 0; 161 } 162 } 163 // This pass just sets some internal MachineMemOperand flags. It can't really 164 // invalidate anything. 165 return false; 166} 167