ARMBaseInstrInfo.cpp revision 3eca98bb3ab1ec27ab8763298c416d282cdaa261
1//===- ARMBaseInstrInfo.cpp - ARM Instruction Information -------*- C++ -*-===//
2//
3//                     The LLVM Compiler Infrastructure
4//
5// This file is distributed under the University of Illinois Open Source
6// License. See LICENSE.TXT for details.
7//
8//===----------------------------------------------------------------------===//
9//
10// This file contains the Base ARM implementation of the TargetInstrInfo class.
11//
12//===----------------------------------------------------------------------===//
13
14#include "ARMBaseInstrInfo.h"
15#include "ARM.h"
16#include "ARMAddressingModes.h"
17#include "ARMConstantPoolValue.h"
18#include "ARMGenInstrInfo.inc"
19#include "ARMMachineFunctionInfo.h"
20#include "ARMRegisterInfo.h"
21#include "llvm/Constants.h"
22#include "llvm/Function.h"
23#include "llvm/GlobalValue.h"
24#include "llvm/ADT/STLExtras.h"
25#include "llvm/CodeGen/LiveVariables.h"
26#include "llvm/CodeGen/MachineConstantPool.h"
27#include "llvm/CodeGen/MachineFrameInfo.h"
28#include "llvm/CodeGen/MachineInstrBuilder.h"
29#include "llvm/CodeGen/MachineJumpTableInfo.h"
30#include "llvm/CodeGen/MachineMemOperand.h"
31#include "llvm/CodeGen/MachineRegisterInfo.h"
32#include "llvm/CodeGen/PseudoSourceValue.h"
33#include "llvm/MC/MCAsmInfo.h"
34#include "llvm/Support/CommandLine.h"
35#include "llvm/Support/Debug.h"
36#include "llvm/Support/ErrorHandling.h"
37using namespace llvm;
38
39static cl::opt<bool>
40EnableARM3Addr("enable-arm-3-addr-conv", cl::Hidden,
41               cl::desc("Enable ARM 2-addr to 3-addr conv"));
42
43ARMBaseInstrInfo::ARMBaseInstrInfo(const ARMSubtarget& STI)
44  : TargetInstrInfoImpl(ARMInsts, array_lengthof(ARMInsts)),
45    Subtarget(STI) {
46}
47
48MachineInstr *
49ARMBaseInstrInfo::convertToThreeAddress(MachineFunction::iterator &MFI,
50                                        MachineBasicBlock::iterator &MBBI,
51                                        LiveVariables *LV) const {
52  // FIXME: Thumb2 support.
53
54  if (!EnableARM3Addr)
55    return NULL;
56
57  MachineInstr *MI = MBBI;
58  MachineFunction &MF = *MI->getParent()->getParent();
59  uint64_t TSFlags = MI->getDesc().TSFlags;
60  bool isPre = false;
61  switch ((TSFlags & ARMII::IndexModeMask) >> ARMII::IndexModeShift) {
62  default: return NULL;
63  case ARMII::IndexModePre:
64    isPre = true;
65    break;
66  case ARMII::IndexModePost:
67    break;
68  }
69
70  // Try splitting an indexed load/store to an un-indexed one plus an add/sub
71  // operation.
72  unsigned MemOpc = getUnindexedOpcode(MI->getOpcode());
73  if (MemOpc == 0)
74    return NULL;
75
76  MachineInstr *UpdateMI = NULL;
77  MachineInstr *MemMI = NULL;
78  unsigned AddrMode = (TSFlags & ARMII::AddrModeMask);
79  const TargetInstrDesc &TID = MI->getDesc();
80  unsigned NumOps = TID.getNumOperands();
81  bool isLoad = !TID.mayStore();
82  const MachineOperand &WB = isLoad ? MI->getOperand(1) : MI->getOperand(0);
83  const MachineOperand &Base = MI->getOperand(2);
84  const MachineOperand &Offset = MI->getOperand(NumOps-3);
85  unsigned WBReg = WB.getReg();
86  unsigned BaseReg = Base.getReg();
87  unsigned OffReg = Offset.getReg();
88  unsigned OffImm = MI->getOperand(NumOps-2).getImm();
89  ARMCC::CondCodes Pred = (ARMCC::CondCodes)MI->getOperand(NumOps-1).getImm();
90  switch (AddrMode) {
91  default:
92    assert(false && "Unknown indexed op!");
93    return NULL;
94  case ARMII::AddrMode2: {
95    bool isSub = ARM_AM::getAM2Op(OffImm) == ARM_AM::sub;
96    unsigned Amt = ARM_AM::getAM2Offset(OffImm);
97    if (OffReg == 0) {
98      if (ARM_AM::getSOImmVal(Amt) == -1)
99        // Can't encode it in a so_imm operand. This transformation will
100        // add more than 1 instruction. Abandon!
101        return NULL;
102      UpdateMI = BuildMI(MF, MI->getDebugLoc(),
103                         get(isSub ? ARM::SUBri : ARM::ADDri), WBReg)
104        .addReg(BaseReg).addImm(Amt)
105        .addImm(Pred).addReg(0).addReg(0);
106    } else if (Amt != 0) {
107      ARM_AM::ShiftOpc ShOpc = ARM_AM::getAM2ShiftOpc(OffImm);
108      unsigned SOOpc = ARM_AM::getSORegOpc(ShOpc, Amt);
109      UpdateMI = BuildMI(MF, MI->getDebugLoc(),
110                         get(isSub ? ARM::SUBrs : ARM::ADDrs), WBReg)
111        .addReg(BaseReg).addReg(OffReg).addReg(0).addImm(SOOpc)
112        .addImm(Pred).addReg(0).addReg(0);
113    } else
114      UpdateMI = BuildMI(MF, MI->getDebugLoc(),
115                         get(isSub ? ARM::SUBrr : ARM::ADDrr), WBReg)
116        .addReg(BaseReg).addReg(OffReg)
117        .addImm(Pred).addReg(0).addReg(0);
118    break;
119  }
120  case ARMII::AddrMode3 : {
121    bool isSub = ARM_AM::getAM3Op(OffImm) == ARM_AM::sub;
122    unsigned Amt = ARM_AM::getAM3Offset(OffImm);
123    if (OffReg == 0)
124      // Immediate is 8-bits. It's guaranteed to fit in a so_imm operand.
125      UpdateMI = BuildMI(MF, MI->getDebugLoc(),
126                         get(isSub ? ARM::SUBri : ARM::ADDri), WBReg)
127        .addReg(BaseReg).addImm(Amt)
128        .addImm(Pred).addReg(0).addReg(0);
129    else
130      UpdateMI = BuildMI(MF, MI->getDebugLoc(),
131                         get(isSub ? ARM::SUBrr : ARM::ADDrr), WBReg)
132        .addReg(BaseReg).addReg(OffReg)
133        .addImm(Pred).addReg(0).addReg(0);
134    break;
135  }
136  }
137
138  std::vector<MachineInstr*> NewMIs;
139  if (isPre) {
140    if (isLoad)
141      MemMI = BuildMI(MF, MI->getDebugLoc(),
142                      get(MemOpc), MI->getOperand(0).getReg())
143        .addReg(WBReg).addReg(0).addImm(0).addImm(Pred);
144    else
145      MemMI = BuildMI(MF, MI->getDebugLoc(),
146                      get(MemOpc)).addReg(MI->getOperand(1).getReg())
147        .addReg(WBReg).addReg(0).addImm(0).addImm(Pred);
148    NewMIs.push_back(MemMI);
149    NewMIs.push_back(UpdateMI);
150  } else {
151    if (isLoad)
152      MemMI = BuildMI(MF, MI->getDebugLoc(),
153                      get(MemOpc), MI->getOperand(0).getReg())
154        .addReg(BaseReg).addReg(0).addImm(0).addImm(Pred);
155    else
156      MemMI = BuildMI(MF, MI->getDebugLoc(),
157                      get(MemOpc)).addReg(MI->getOperand(1).getReg())
158        .addReg(BaseReg).addReg(0).addImm(0).addImm(Pred);
159    if (WB.isDead())
160      UpdateMI->getOperand(0).setIsDead();
161    NewMIs.push_back(UpdateMI);
162    NewMIs.push_back(MemMI);
163  }
164
165  // Transfer LiveVariables states, kill / dead info.
166  if (LV) {
167    for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) {
168      MachineOperand &MO = MI->getOperand(i);
169      if (MO.isReg() && MO.getReg() &&
170          TargetRegisterInfo::isVirtualRegister(MO.getReg())) {
171        unsigned Reg = MO.getReg();
172
173        LiveVariables::VarInfo &VI = LV->getVarInfo(Reg);
174        if (MO.isDef()) {
175          MachineInstr *NewMI = (Reg == WBReg) ? UpdateMI : MemMI;
176          if (MO.isDead())
177            LV->addVirtualRegisterDead(Reg, NewMI);
178        }
179        if (MO.isUse() && MO.isKill()) {
180          for (unsigned j = 0; j < 2; ++j) {
181            // Look at the two new MI's in reverse order.
182            MachineInstr *NewMI = NewMIs[j];
183            if (!NewMI->readsRegister(Reg))
184              continue;
185            LV->addVirtualRegisterKilled(Reg, NewMI);
186            if (VI.removeKill(MI))
187              VI.Kills.push_back(NewMI);
188            break;
189          }
190        }
191      }
192    }
193  }
194
195  MFI->insert(MBBI, NewMIs[1]);
196  MFI->insert(MBBI, NewMIs[0]);
197  return NewMIs[0];
198}
199
200bool
201ARMBaseInstrInfo::spillCalleeSavedRegisters(MachineBasicBlock &MBB,
202                                        MachineBasicBlock::iterator MI,
203                                        const std::vector<CalleeSavedInfo> &CSI,
204                                        const TargetRegisterInfo *TRI) const {
205  if (CSI.empty())
206    return false;
207
208  DebugLoc DL;
209  if (MI != MBB.end()) DL = MI->getDebugLoc();
210
211  for (unsigned i = 0, e = CSI.size(); i != e; ++i) {
212    unsigned Reg = CSI[i].getReg();
213    bool isKill = true;
214
215    // Add the callee-saved register as live-in unless it's LR and
216    // @llvm.returnaddress is called. If LR is returned for @llvm.returnaddress
217    // then it's already added to the function and entry block live-in sets.
218    if (Reg == ARM::LR) {
219      MachineFunction &MF = *MBB.getParent();
220      if (MF.getFrameInfo()->isReturnAddressTaken() &&
221          MF.getRegInfo().isLiveIn(Reg))
222        isKill = false;
223    }
224
225    if (isKill)
226      MBB.addLiveIn(Reg);
227
228    // Insert the spill to the stack frame. The register is killed at the spill
229    //
230    const TargetRegisterClass *RC = TRI->getMinimalPhysRegClass(Reg);
231    storeRegToStackSlot(MBB, MI, Reg, isKill,
232                        CSI[i].getFrameIdx(), RC, TRI);
233  }
234  return true;
235}
236
237// Branch analysis.
238bool
239ARMBaseInstrInfo::AnalyzeBranch(MachineBasicBlock &MBB,MachineBasicBlock *&TBB,
240                                MachineBasicBlock *&FBB,
241                                SmallVectorImpl<MachineOperand> &Cond,
242                                bool AllowModify) const {
243  // If the block has no terminators, it just falls into the block after it.
244  MachineBasicBlock::iterator I = MBB.end();
245  if (I == MBB.begin())
246    return false;
247  --I;
248  while (I->isDebugValue()) {
249    if (I == MBB.begin())
250      return false;
251    --I;
252  }
253  if (!isUnpredicatedTerminator(I))
254    return false;
255
256  // Get the last instruction in the block.
257  MachineInstr *LastInst = I;
258
259  // If there is only one terminator instruction, process it.
260  unsigned LastOpc = LastInst->getOpcode();
261  if (I == MBB.begin() || !isUnpredicatedTerminator(--I)) {
262    if (isUncondBranchOpcode(LastOpc)) {
263      TBB = LastInst->getOperand(0).getMBB();
264      return false;
265    }
266    if (isCondBranchOpcode(LastOpc)) {
267      // Block ends with fall-through condbranch.
268      TBB = LastInst->getOperand(0).getMBB();
269      Cond.push_back(LastInst->getOperand(1));
270      Cond.push_back(LastInst->getOperand(2));
271      return false;
272    }
273    return true;  // Can't handle indirect branch.
274  }
275
276  // Get the instruction before it if it is a terminator.
277  MachineInstr *SecondLastInst = I;
278
279  // If there are three terminators, we don't know what sort of block this is.
280  if (SecondLastInst && I != MBB.begin() && isUnpredicatedTerminator(--I))
281    return true;
282
283  // If the block ends with a B and a Bcc, handle it.
284  unsigned SecondLastOpc = SecondLastInst->getOpcode();
285  if (isCondBranchOpcode(SecondLastOpc) && isUncondBranchOpcode(LastOpc)) {
286    TBB =  SecondLastInst->getOperand(0).getMBB();
287    Cond.push_back(SecondLastInst->getOperand(1));
288    Cond.push_back(SecondLastInst->getOperand(2));
289    FBB = LastInst->getOperand(0).getMBB();
290    return false;
291  }
292
293  // If the block ends with two unconditional branches, handle it.  The second
294  // one is not executed, so remove it.
295  if (isUncondBranchOpcode(SecondLastOpc) && isUncondBranchOpcode(LastOpc)) {
296    TBB = SecondLastInst->getOperand(0).getMBB();
297    I = LastInst;
298    if (AllowModify)
299      I->eraseFromParent();
300    return false;
301  }
302
303  // ...likewise if it ends with a branch table followed by an unconditional
304  // branch. The branch folder can create these, and we must get rid of them for
305  // correctness of Thumb constant islands.
306  if ((isJumpTableBranchOpcode(SecondLastOpc) ||
307       isIndirectBranchOpcode(SecondLastOpc)) &&
308      isUncondBranchOpcode(LastOpc)) {
309    I = LastInst;
310    if (AllowModify)
311      I->eraseFromParent();
312    return true;
313  }
314
315  // Otherwise, can't handle this.
316  return true;
317}
318
319
320unsigned ARMBaseInstrInfo::RemoveBranch(MachineBasicBlock &MBB) const {
321  MachineBasicBlock::iterator I = MBB.end();
322  if (I == MBB.begin()) return 0;
323  --I;
324  while (I->isDebugValue()) {
325    if (I == MBB.begin())
326      return 0;
327    --I;
328  }
329  if (!isUncondBranchOpcode(I->getOpcode()) &&
330      !isCondBranchOpcode(I->getOpcode()))
331    return 0;
332
333  // Remove the branch.
334  I->eraseFromParent();
335
336  I = MBB.end();
337
338  if (I == MBB.begin()) return 1;
339  --I;
340  if (!isCondBranchOpcode(I->getOpcode()))
341    return 1;
342
343  // Remove the branch.
344  I->eraseFromParent();
345  return 2;
346}
347
348unsigned
349ARMBaseInstrInfo::InsertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB,
350                            MachineBasicBlock *FBB,
351                            const SmallVectorImpl<MachineOperand> &Cond) const {
352  // FIXME this should probably have a DebugLoc argument
353  DebugLoc dl;
354
355  ARMFunctionInfo *AFI = MBB.getParent()->getInfo<ARMFunctionInfo>();
356  int BOpc   = !AFI->isThumbFunction()
357    ? ARM::B : (AFI->isThumb2Function() ? ARM::t2B : ARM::tB);
358  int BccOpc = !AFI->isThumbFunction()
359    ? ARM::Bcc : (AFI->isThumb2Function() ? ARM::t2Bcc : ARM::tBcc);
360
361  // Shouldn't be a fall through.
362  assert(TBB && "InsertBranch must not be told to insert a fallthrough");
363  assert((Cond.size() == 2 || Cond.size() == 0) &&
364         "ARM branch conditions have two components!");
365
366  if (FBB == 0) {
367    if (Cond.empty()) // Unconditional branch?
368      BuildMI(&MBB, dl, get(BOpc)).addMBB(TBB);
369    else
370      BuildMI(&MBB, dl, get(BccOpc)).addMBB(TBB)
371        .addImm(Cond[0].getImm()).addReg(Cond[1].getReg());
372    return 1;
373  }
374
375  // Two-way conditional branch.
376  BuildMI(&MBB, dl, get(BccOpc)).addMBB(TBB)
377    .addImm(Cond[0].getImm()).addReg(Cond[1].getReg());
378  BuildMI(&MBB, dl, get(BOpc)).addMBB(FBB);
379  return 2;
380}
381
382bool ARMBaseInstrInfo::
383ReverseBranchCondition(SmallVectorImpl<MachineOperand> &Cond) const {
384  ARMCC::CondCodes CC = (ARMCC::CondCodes)(int)Cond[0].getImm();
385  Cond[0].setImm(ARMCC::getOppositeCondition(CC));
386  return false;
387}
388
389bool ARMBaseInstrInfo::
390PredicateInstruction(MachineInstr *MI,
391                     const SmallVectorImpl<MachineOperand> &Pred) const {
392  unsigned Opc = MI->getOpcode();
393  if (isUncondBranchOpcode(Opc)) {
394    MI->setDesc(get(getMatchingCondBranchOpcode(Opc)));
395    MI->addOperand(MachineOperand::CreateImm(Pred[0].getImm()));
396    MI->addOperand(MachineOperand::CreateReg(Pred[1].getReg(), false));
397    return true;
398  }
399
400  int PIdx = MI->findFirstPredOperandIdx();
401  if (PIdx != -1) {
402    MachineOperand &PMO = MI->getOperand(PIdx);
403    PMO.setImm(Pred[0].getImm());
404    MI->getOperand(PIdx+1).setReg(Pred[1].getReg());
405    return true;
406  }
407  return false;
408}
409
410bool ARMBaseInstrInfo::
411SubsumesPredicate(const SmallVectorImpl<MachineOperand> &Pred1,
412                  const SmallVectorImpl<MachineOperand> &Pred2) const {
413  if (Pred1.size() > 2 || Pred2.size() > 2)
414    return false;
415
416  ARMCC::CondCodes CC1 = (ARMCC::CondCodes)Pred1[0].getImm();
417  ARMCC::CondCodes CC2 = (ARMCC::CondCodes)Pred2[0].getImm();
418  if (CC1 == CC2)
419    return true;
420
421  switch (CC1) {
422  default:
423    return false;
424  case ARMCC::AL:
425    return true;
426  case ARMCC::HS:
427    return CC2 == ARMCC::HI;
428  case ARMCC::LS:
429    return CC2 == ARMCC::LO || CC2 == ARMCC::EQ;
430  case ARMCC::GE:
431    return CC2 == ARMCC::GT;
432  case ARMCC::LE:
433    return CC2 == ARMCC::LT;
434  }
435}
436
437bool ARMBaseInstrInfo::DefinesPredicate(MachineInstr *MI,
438                                    std::vector<MachineOperand> &Pred) const {
439  // FIXME: This confuses implicit_def with optional CPSR def.
440  const TargetInstrDesc &TID = MI->getDesc();
441  if (!TID.getImplicitDefs() && !TID.hasOptionalDef())
442    return false;
443
444  bool Found = false;
445  for (unsigned i = 0, e = MI->getNumOperands(); i != e; ++i) {
446    const MachineOperand &MO = MI->getOperand(i);
447    if (MO.isReg() && MO.getReg() == ARM::CPSR) {
448      Pred.push_back(MO);
449      Found = true;
450    }
451  }
452
453  return Found;
454}
455
456/// isPredicable - Return true if the specified instruction can be predicated.
457/// By default, this returns true for every instruction with a
458/// PredicateOperand.
459bool ARMBaseInstrInfo::isPredicable(MachineInstr *MI) const {
460  const TargetInstrDesc &TID = MI->getDesc();
461  if (!TID.isPredicable())
462    return false;
463
464  if ((TID.TSFlags & ARMII::DomainMask) == ARMII::DomainNEON) {
465    ARMFunctionInfo *AFI =
466      MI->getParent()->getParent()->getInfo<ARMFunctionInfo>();
467    return AFI->isThumb2Function();
468  }
469  return true;
470}
471
472/// FIXME: Works around a gcc miscompilation with -fstrict-aliasing.
473DISABLE_INLINE
474static unsigned getNumJTEntries(const std::vector<MachineJumpTableEntry> &JT,
475                                unsigned JTI);
476static unsigned getNumJTEntries(const std::vector<MachineJumpTableEntry> &JT,
477                                unsigned JTI) {
478  assert(JTI < JT.size());
479  return JT[JTI].MBBs.size();
480}
481
482/// GetInstSize - Return the size of the specified MachineInstr.
483///
484unsigned ARMBaseInstrInfo::GetInstSizeInBytes(const MachineInstr *MI) const {
485  const MachineBasicBlock &MBB = *MI->getParent();
486  const MachineFunction *MF = MBB.getParent();
487  const MCAsmInfo *MAI = MF->getTarget().getMCAsmInfo();
488
489  // Basic size info comes from the TSFlags field.
490  const TargetInstrDesc &TID = MI->getDesc();
491  uint64_t TSFlags = TID.TSFlags;
492
493  unsigned Opc = MI->getOpcode();
494  switch ((TSFlags & ARMII::SizeMask) >> ARMII::SizeShift) {
495  default: {
496    // If this machine instr is an inline asm, measure it.
497    if (MI->getOpcode() == ARM::INLINEASM)
498      return getInlineAsmLength(MI->getOperand(0).getSymbolName(), *MAI);
499    if (MI->isLabel())
500      return 0;
501    switch (Opc) {
502    default:
503      llvm_unreachable("Unknown or unset size field for instr!");
504    case TargetOpcode::IMPLICIT_DEF:
505    case TargetOpcode::KILL:
506    case TargetOpcode::DBG_LABEL:
507    case TargetOpcode::EH_LABEL:
508    case TargetOpcode::DBG_VALUE:
509      return 0;
510    }
511    break;
512  }
513  case ARMII::Size8Bytes: return 8;          // ARM instruction x 2.
514  case ARMII::Size4Bytes: return 4;          // ARM / Thumb2 instruction.
515  case ARMII::Size2Bytes: return 2;          // Thumb1 instruction.
516  case ARMII::SizeSpecial: {
517    switch (Opc) {
518    case ARM::CONSTPOOL_ENTRY:
519      // If this machine instr is a constant pool entry, its size is recorded as
520      // operand #2.
521      return MI->getOperand(2).getImm();
522    case ARM::Int_eh_sjlj_longjmp:
523      return 16;
524    case ARM::tInt_eh_sjlj_longjmp:
525      return 10;
526    case ARM::Int_eh_sjlj_setjmp:
527    case ARM::Int_eh_sjlj_setjmp_nofp:
528      return 20;
529    case ARM::tInt_eh_sjlj_setjmp:
530    case ARM::t2Int_eh_sjlj_setjmp:
531    case ARM::t2Int_eh_sjlj_setjmp_nofp:
532      return 12;
533    case ARM::BR_JTr:
534    case ARM::BR_JTm:
535    case ARM::BR_JTadd:
536    case ARM::tBR_JTr:
537    case ARM::t2BR_JT:
538    case ARM::t2TBB:
539    case ARM::t2TBH: {
540      // These are jumptable branches, i.e. a branch followed by an inlined
541      // jumptable. The size is 4 + 4 * number of entries. For TBB, each
542      // entry is one byte; TBH two byte each.
543      unsigned EntrySize = (Opc == ARM::t2TBB)
544        ? 1 : ((Opc == ARM::t2TBH) ? 2 : 4);
545      unsigned NumOps = TID.getNumOperands();
546      MachineOperand JTOP =
547        MI->getOperand(NumOps - (TID.isPredicable() ? 3 : 2));
548      unsigned JTI = JTOP.getIndex();
549      const MachineJumpTableInfo *MJTI = MF->getJumpTableInfo();
550      assert(MJTI != 0);
551      const std::vector<MachineJumpTableEntry> &JT = MJTI->getJumpTables();
552      assert(JTI < JT.size());
553      // Thumb instructions are 2 byte aligned, but JT entries are 4 byte
554      // 4 aligned. The assembler / linker may add 2 byte padding just before
555      // the JT entries.  The size does not include this padding; the
556      // constant islands pass does separate bookkeeping for it.
557      // FIXME: If we know the size of the function is less than (1 << 16) *2
558      // bytes, we can use 16-bit entries instead. Then there won't be an
559      // alignment issue.
560      unsigned InstSize = (Opc == ARM::tBR_JTr || Opc == ARM::t2BR_JT) ? 2 : 4;
561      unsigned NumEntries = getNumJTEntries(JT, JTI);
562      if (Opc == ARM::t2TBB && (NumEntries & 1))
563        // Make sure the instruction that follows TBB is 2-byte aligned.
564        // FIXME: Constant island pass should insert an "ALIGN" instruction
565        // instead.
566        ++NumEntries;
567      return NumEntries * EntrySize + InstSize;
568    }
569    default:
570      // Otherwise, pseudo-instruction sizes are zero.
571      return 0;
572    }
573  }
574  }
575  return 0; // Not reached
576}
577
578/// Return true if the instruction is a register to register move and
579/// leave the source and dest operands in the passed parameters.
580///
581bool
582ARMBaseInstrInfo::isMoveInstr(const MachineInstr &MI,
583                              unsigned &SrcReg, unsigned &DstReg,
584                              unsigned& SrcSubIdx, unsigned& DstSubIdx) const {
585  switch (MI.getOpcode()) {
586  default: break;
587  case ARM::VMOVS:
588  case ARM::VMOVD:
589  case ARM::VMOVDneon:
590  case ARM::VMOVQ:
591  case ARM::VMOVQQ : {
592    SrcReg = MI.getOperand(1).getReg();
593    DstReg = MI.getOperand(0).getReg();
594    SrcSubIdx = MI.getOperand(1).getSubReg();
595    DstSubIdx = MI.getOperand(0).getSubReg();
596    return true;
597  }
598  case ARM::MOVr:
599  case ARM::tMOVr:
600  case ARM::tMOVgpr2tgpr:
601  case ARM::tMOVtgpr2gpr:
602  case ARM::tMOVgpr2gpr:
603  case ARM::t2MOVr: {
604    assert(MI.getDesc().getNumOperands() >= 2 &&
605           MI.getOperand(0).isReg() &&
606           MI.getOperand(1).isReg() &&
607           "Invalid ARM MOV instruction");
608    SrcReg = MI.getOperand(1).getReg();
609    DstReg = MI.getOperand(0).getReg();
610    SrcSubIdx = MI.getOperand(1).getSubReg();
611    DstSubIdx = MI.getOperand(0).getSubReg();
612    return true;
613  }
614  }
615
616  return false;
617}
618
619unsigned
620ARMBaseInstrInfo::isLoadFromStackSlot(const MachineInstr *MI,
621                                      int &FrameIndex) const {
622  switch (MI->getOpcode()) {
623  default: break;
624  case ARM::LDR:
625  case ARM::t2LDRs:  // FIXME: don't use t2LDRs to access frame.
626    if (MI->getOperand(1).isFI() &&
627        MI->getOperand(2).isReg() &&
628        MI->getOperand(3).isImm() &&
629        MI->getOperand(2).getReg() == 0 &&
630        MI->getOperand(3).getImm() == 0) {
631      FrameIndex = MI->getOperand(1).getIndex();
632      return MI->getOperand(0).getReg();
633    }
634    break;
635  case ARM::t2LDRi12:
636  case ARM::tRestore:
637    if (MI->getOperand(1).isFI() &&
638        MI->getOperand(2).isImm() &&
639        MI->getOperand(2).getImm() == 0) {
640      FrameIndex = MI->getOperand(1).getIndex();
641      return MI->getOperand(0).getReg();
642    }
643    break;
644  case ARM::VLDRD:
645  case ARM::VLDRS:
646    if (MI->getOperand(1).isFI() &&
647        MI->getOperand(2).isImm() &&
648        MI->getOperand(2).getImm() == 0) {
649      FrameIndex = MI->getOperand(1).getIndex();
650      return MI->getOperand(0).getReg();
651    }
652    break;
653  }
654
655  return 0;
656}
657
658unsigned
659ARMBaseInstrInfo::isStoreToStackSlot(const MachineInstr *MI,
660                                     int &FrameIndex) const {
661  switch (MI->getOpcode()) {
662  default: break;
663  case ARM::STR:
664  case ARM::t2STRs: // FIXME: don't use t2STRs to access frame.
665    if (MI->getOperand(1).isFI() &&
666        MI->getOperand(2).isReg() &&
667        MI->getOperand(3).isImm() &&
668        MI->getOperand(2).getReg() == 0 &&
669        MI->getOperand(3).getImm() == 0) {
670      FrameIndex = MI->getOperand(1).getIndex();
671      return MI->getOperand(0).getReg();
672    }
673    break;
674  case ARM::t2STRi12:
675  case ARM::tSpill:
676    if (MI->getOperand(1).isFI() &&
677        MI->getOperand(2).isImm() &&
678        MI->getOperand(2).getImm() == 0) {
679      FrameIndex = MI->getOperand(1).getIndex();
680      return MI->getOperand(0).getReg();
681    }
682    break;
683  case ARM::VSTRD:
684  case ARM::VSTRS:
685    if (MI->getOperand(1).isFI() &&
686        MI->getOperand(2).isImm() &&
687        MI->getOperand(2).getImm() == 0) {
688      FrameIndex = MI->getOperand(1).getIndex();
689      return MI->getOperand(0).getReg();
690    }
691    break;
692  }
693
694  return 0;
695}
696
697bool
698ARMBaseInstrInfo::copyRegToReg(MachineBasicBlock &MBB,
699                               MachineBasicBlock::iterator I,
700                               unsigned DestReg, unsigned SrcReg,
701                               const TargetRegisterClass *DestRC,
702                               const TargetRegisterClass *SrcRC,
703                               DebugLoc DL) const {
704  // tGPR is used sometimes in ARM instructions that need to avoid using
705  // certain registers.  Just treat it as GPR here.
706  if (DestRC == ARM::tGPRRegisterClass)
707    DestRC = ARM::GPRRegisterClass;
708  if (SrcRC == ARM::tGPRRegisterClass)
709    SrcRC = ARM::GPRRegisterClass;
710
711  // Allow DPR / DPR_VFP2 / DPR_8 cross-class copies.
712  if (DestRC == ARM::DPR_8RegisterClass)
713    DestRC = ARM::DPR_VFP2RegisterClass;
714  if (SrcRC == ARM::DPR_8RegisterClass)
715    SrcRC = ARM::DPR_VFP2RegisterClass;
716
717  // Allow QPR / QPR_VFP2 / QPR_8 cross-class copies.
718  if (DestRC == ARM::QPR_VFP2RegisterClass ||
719      DestRC == ARM::QPR_8RegisterClass)
720    DestRC = ARM::QPRRegisterClass;
721  if (SrcRC == ARM::QPR_VFP2RegisterClass ||
722      SrcRC == ARM::QPR_8RegisterClass)
723    SrcRC = ARM::QPRRegisterClass;
724
725  // Allow QQPR / QQPR_VFP2 cross-class copies.
726  if (DestRC == ARM::QQPR_VFP2RegisterClass)
727    DestRC = ARM::QQPRRegisterClass;
728  if (SrcRC == ARM::QQPR_VFP2RegisterClass)
729    SrcRC = ARM::QQPRRegisterClass;
730
731  // Disallow copies of unequal sizes.
732  if (DestRC != SrcRC && DestRC->getSize() != SrcRC->getSize())
733    return false;
734
735  if (DestRC == ARM::GPRRegisterClass) {
736    if (SrcRC == ARM::SPRRegisterClass)
737      AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VMOVRS), DestReg)
738                     .addReg(SrcReg));
739    else
740      AddDefaultCC(AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::MOVr),
741                                          DestReg).addReg(SrcReg)));
742  } else {
743    unsigned Opc;
744
745    if (DestRC == ARM::SPRRegisterClass)
746      Opc = (SrcRC == ARM::GPRRegisterClass ? ARM::VMOVSR : ARM::VMOVS);
747    else if (DestRC == ARM::DPRRegisterClass)
748      Opc = ARM::VMOVD;
749    else if (DestRC == ARM::DPR_VFP2RegisterClass ||
750             SrcRC == ARM::DPR_VFP2RegisterClass)
751      // Always use neon reg-reg move if source or dest is NEON-only regclass.
752      Opc = ARM::VMOVDneon;
753    else if (DestRC == ARM::QPRRegisterClass)
754      Opc = ARM::VMOVQ;
755    else if (DestRC == ARM::QQPRRegisterClass)
756      Opc = ARM::VMOVQQ;
757    else if (DestRC == ARM::QQQQPRRegisterClass)
758      Opc = ARM::VMOVQQQQ;
759    else
760      return false;
761
762    AddDefaultPred(BuildMI(MBB, I, DL, get(Opc), DestReg).addReg(SrcReg));
763  }
764
765  return true;
766}
767
768static const
769MachineInstrBuilder &AddDReg(MachineInstrBuilder &MIB,
770                             unsigned Reg, unsigned SubIdx, unsigned State,
771                             const TargetRegisterInfo *TRI) {
772  if (!SubIdx)
773    return MIB.addReg(Reg, State);
774
775  if (TargetRegisterInfo::isPhysicalRegister(Reg))
776    return MIB.addReg(TRI->getSubReg(Reg, SubIdx), State);
777  return MIB.addReg(Reg, State, SubIdx);
778}
779
780void ARMBaseInstrInfo::
781storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
782                    unsigned SrcReg, bool isKill, int FI,
783                    const TargetRegisterClass *RC,
784                    const TargetRegisterInfo *TRI) const {
785  DebugLoc DL;
786  if (I != MBB.end()) DL = I->getDebugLoc();
787  MachineFunction &MF = *MBB.getParent();
788  MachineFrameInfo &MFI = *MF.getFrameInfo();
789  unsigned Align = MFI.getObjectAlignment(FI);
790
791  MachineMemOperand *MMO =
792    MF.getMachineMemOperand(PseudoSourceValue::getFixedStack(FI),
793                            MachineMemOperand::MOStore, 0,
794                            MFI.getObjectSize(FI),
795                            Align);
796
797  // tGPR is used sometimes in ARM instructions that need to avoid using
798  // certain registers.  Just treat it as GPR here.
799  if (RC == ARM::tGPRRegisterClass)
800    RC = ARM::GPRRegisterClass;
801
802  if (RC == ARM::GPRRegisterClass) {
803    AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::STR))
804                   .addReg(SrcReg, getKillRegState(isKill))
805                   .addFrameIndex(FI).addReg(0).addImm(0).addMemOperand(MMO));
806  } else if (RC == ARM::SPRRegisterClass) {
807    AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VSTRS))
808                   .addReg(SrcReg, getKillRegState(isKill))
809                   .addFrameIndex(FI).addImm(0).addMemOperand(MMO));
810  } else if (RC == ARM::DPRRegisterClass ||
811             RC == ARM::DPR_VFP2RegisterClass ||
812             RC == ARM::DPR_8RegisterClass) {
813    AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VSTRD))
814                   .addReg(SrcReg, getKillRegState(isKill))
815                   .addFrameIndex(FI).addImm(0).addMemOperand(MMO));
816  } else if (RC == ARM::QPRRegisterClass ||
817             RC == ARM::QPR_VFP2RegisterClass ||
818             RC == ARM::QPR_8RegisterClass) {
819    // FIXME: Neon instructions should support predicates
820    if (Align >= 16 && getRegisterInfo().canRealignStack(MF)) {
821      AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VST1q))
822                     .addFrameIndex(FI).addImm(128)
823                     .addReg(SrcReg, getKillRegState(isKill))
824                     .addMemOperand(MMO));
825    } else {
826      AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VSTMQ))
827                     .addReg(SrcReg, getKillRegState(isKill))
828                     .addFrameIndex(FI)
829                     .addImm(ARM_AM::getAM5Opc(ARM_AM::ia, 4))
830                     .addMemOperand(MMO));
831    }
832  } else if (RC == ARM::QQPRRegisterClass || RC == ARM::QQPR_VFP2RegisterClass){
833    if (Align >= 16 && getRegisterInfo().canRealignStack(MF)) {
834      // FIXME: It's possible to only store part of the QQ register if the
835      // spilled def has a sub-register index.
836      MachineInstrBuilder MIB = BuildMI(MBB, I, DL, get(ARM::VST2q32))
837        .addFrameIndex(FI).addImm(128);
838      MIB = AddDReg(MIB, SrcReg, ARM::dsub_0, getKillRegState(isKill), TRI);
839      MIB = AddDReg(MIB, SrcReg, ARM::dsub_1, 0, TRI);
840      MIB = AddDReg(MIB, SrcReg, ARM::dsub_2, 0, TRI);
841      MIB = AddDReg(MIB, SrcReg, ARM::dsub_3, 0, TRI);
842      AddDefaultPred(MIB.addMemOperand(MMO));
843    } else {
844      MachineInstrBuilder MIB =
845        AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VSTMD))
846                       .addFrameIndex(FI)
847                       .addImm(ARM_AM::getAM5Opc(ARM_AM::ia, 4)))
848        .addMemOperand(MMO);
849      MIB = AddDReg(MIB, SrcReg, ARM::dsub_0, getKillRegState(isKill), TRI);
850      MIB = AddDReg(MIB, SrcReg, ARM::dsub_1, 0, TRI);
851      MIB = AddDReg(MIB, SrcReg, ARM::dsub_2, 0, TRI);
852            AddDReg(MIB, SrcReg, ARM::dsub_3, 0, TRI);
853    }
854  } else {
855    assert(RC == ARM::QQQQPRRegisterClass && "Unknown regclass!");
856    MachineInstrBuilder MIB =
857      AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VSTMD))
858                     .addFrameIndex(FI)
859                     .addImm(ARM_AM::getAM5Opc(ARM_AM::ia, 4)))
860      .addMemOperand(MMO);
861    MIB = AddDReg(MIB, SrcReg, ARM::dsub_0, getKillRegState(isKill), TRI);
862    MIB = AddDReg(MIB, SrcReg, ARM::dsub_1, 0, TRI);
863    MIB = AddDReg(MIB, SrcReg, ARM::dsub_2, 0, TRI);
864    MIB = AddDReg(MIB, SrcReg, ARM::dsub_3, 0, TRI);
865    MIB = AddDReg(MIB, SrcReg, ARM::dsub_4, 0, TRI);
866    MIB = AddDReg(MIB, SrcReg, ARM::dsub_5, 0, TRI);
867    MIB = AddDReg(MIB, SrcReg, ARM::dsub_6, 0, TRI);
868          AddDReg(MIB, SrcReg, ARM::dsub_7, 0, TRI);
869  }
870}
871
872void ARMBaseInstrInfo::
873loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
874                     unsigned DestReg, int FI,
875                     const TargetRegisterClass *RC,
876                     const TargetRegisterInfo *TRI) const {
877  DebugLoc DL;
878  if (I != MBB.end()) DL = I->getDebugLoc();
879  MachineFunction &MF = *MBB.getParent();
880  MachineFrameInfo &MFI = *MF.getFrameInfo();
881  unsigned Align = MFI.getObjectAlignment(FI);
882  MachineMemOperand *MMO =
883    MF.getMachineMemOperand(PseudoSourceValue::getFixedStack(FI),
884                            MachineMemOperand::MOLoad, 0,
885                            MFI.getObjectSize(FI),
886                            Align);
887
888  // tGPR is used sometimes in ARM instructions that need to avoid using
889  // certain registers.  Just treat it as GPR here.
890  if (RC == ARM::tGPRRegisterClass)
891    RC = ARM::GPRRegisterClass;
892
893  if (RC == ARM::GPRRegisterClass) {
894    AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::LDR), DestReg)
895                   .addFrameIndex(FI).addReg(0).addImm(0).addMemOperand(MMO));
896  } else if (RC == ARM::SPRRegisterClass) {
897    AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VLDRS), DestReg)
898                   .addFrameIndex(FI).addImm(0).addMemOperand(MMO));
899  } else if (RC == ARM::DPRRegisterClass ||
900             RC == ARM::DPR_VFP2RegisterClass ||
901             RC == ARM::DPR_8RegisterClass) {
902    AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VLDRD), DestReg)
903                   .addFrameIndex(FI).addImm(0).addMemOperand(MMO));
904  } else if (RC == ARM::QPRRegisterClass ||
905             RC == ARM::QPR_VFP2RegisterClass ||
906             RC == ARM::QPR_8RegisterClass) {
907    if (Align >= 16 && getRegisterInfo().canRealignStack(MF)) {
908      AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VLD1q), DestReg)
909                     .addFrameIndex(FI).addImm(128)
910                     .addMemOperand(MMO));
911    } else {
912      AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VLDMQ), DestReg)
913                     .addFrameIndex(FI)
914                     .addImm(ARM_AM::getAM5Opc(ARM_AM::ia, 4))
915                     .addMemOperand(MMO));
916    }
917  } else if (RC == ARM::QQPRRegisterClass || RC == ARM::QQPR_VFP2RegisterClass){
918    if (Align >= 16 && getRegisterInfo().canRealignStack(MF)) {
919      MachineInstrBuilder MIB = BuildMI(MBB, I, DL, get(ARM::VLD2q32));
920      MIB = AddDReg(MIB, DestReg, ARM::dsub_0, RegState::Define, TRI);
921      MIB = AddDReg(MIB, DestReg, ARM::dsub_1, RegState::Define, TRI);
922      MIB = AddDReg(MIB, DestReg, ARM::dsub_2, RegState::Define, TRI);
923      MIB = AddDReg(MIB, DestReg, ARM::dsub_3, RegState::Define, TRI);
924      AddDefaultPred(MIB.addFrameIndex(FI).addImm(128).addMemOperand(MMO));
925    } else {
926      MachineInstrBuilder MIB =
927        AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VLDMD))
928                       .addFrameIndex(FI)
929                       .addImm(ARM_AM::getAM5Opc(ARM_AM::ia, 4)))
930        .addMemOperand(MMO);
931      MIB = AddDReg(MIB, DestReg, ARM::dsub_0, RegState::Define, TRI);
932      MIB = AddDReg(MIB, DestReg, ARM::dsub_1, RegState::Define, TRI);
933      MIB = AddDReg(MIB, DestReg, ARM::dsub_2, RegState::Define, TRI);
934            AddDReg(MIB, DestReg, ARM::dsub_3, RegState::Define, TRI);
935    }
936  } else {
937    assert(RC == ARM::QQQQPRRegisterClass && "Unknown regclass!");
938      MachineInstrBuilder MIB =
939        AddDefaultPred(BuildMI(MBB, I, DL, get(ARM::VLDMD))
940                       .addFrameIndex(FI)
941                       .addImm(ARM_AM::getAM5Opc(ARM_AM::ia, 4)))
942        .addMemOperand(MMO);
943      MIB = AddDReg(MIB, DestReg, ARM::dsub_0, RegState::Define, TRI);
944      MIB = AddDReg(MIB, DestReg, ARM::dsub_1, RegState::Define, TRI);
945      MIB = AddDReg(MIB, DestReg, ARM::dsub_2, RegState::Define, TRI);
946      MIB = AddDReg(MIB, DestReg, ARM::dsub_3, RegState::Define, TRI);
947      MIB = AddDReg(MIB, DestReg, ARM::dsub_4, RegState::Define, TRI);
948      MIB = AddDReg(MIB, DestReg, ARM::dsub_5, RegState::Define, TRI);
949      MIB = AddDReg(MIB, DestReg, ARM::dsub_6, RegState::Define, TRI);
950            AddDReg(MIB, DestReg, ARM::dsub_7, RegState::Define, TRI);
951  }
952}
953
954MachineInstr*
955ARMBaseInstrInfo::emitFrameIndexDebugValue(MachineFunction &MF,
956                                           int FrameIx, uint64_t Offset,
957                                           const MDNode *MDPtr,
958                                           DebugLoc DL) const {
959  MachineInstrBuilder MIB = BuildMI(MF, DL, get(ARM::DBG_VALUE))
960    .addFrameIndex(FrameIx).addImm(0).addImm(Offset).addMetadata(MDPtr);
961  return &*MIB;
962}
963
964MachineInstr *ARMBaseInstrInfo::
965foldMemoryOperandImpl(MachineFunction &MF, MachineInstr *MI,
966                      const SmallVectorImpl<unsigned> &Ops, int FI) const {
967  if (Ops.size() != 1) return NULL;
968
969  unsigned OpNum = Ops[0];
970  unsigned Opc = MI->getOpcode();
971  MachineInstr *NewMI = NULL;
972  if (Opc == ARM::MOVr || Opc == ARM::t2MOVr) {
973    // If it is updating CPSR, then it cannot be folded.
974    if (MI->getOperand(4).getReg() == ARM::CPSR && !MI->getOperand(4).isDead())
975      return NULL;
976    unsigned Pred = MI->getOperand(2).getImm();
977    unsigned PredReg = MI->getOperand(3).getReg();
978    if (OpNum == 0) { // move -> store
979      unsigned SrcReg = MI->getOperand(1).getReg();
980      unsigned SrcSubReg = MI->getOperand(1).getSubReg();
981      bool isKill = MI->getOperand(1).isKill();
982      bool isUndef = MI->getOperand(1).isUndef();
983      if (Opc == ARM::MOVr)
984        NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::STR))
985          .addReg(SrcReg,
986                  getKillRegState(isKill) | getUndefRegState(isUndef),
987                  SrcSubReg)
988          .addFrameIndex(FI).addReg(0).addImm(0).addImm(Pred).addReg(PredReg);
989      else // ARM::t2MOVr
990        NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::t2STRi12))
991          .addReg(SrcReg,
992                  getKillRegState(isKill) | getUndefRegState(isUndef),
993                  SrcSubReg)
994          .addFrameIndex(FI).addImm(0).addImm(Pred).addReg(PredReg);
995    } else {          // move -> load
996      unsigned DstReg = MI->getOperand(0).getReg();
997      unsigned DstSubReg = MI->getOperand(0).getSubReg();
998      bool isDead = MI->getOperand(0).isDead();
999      bool isUndef = MI->getOperand(0).isUndef();
1000      if (Opc == ARM::MOVr)
1001        NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::LDR))
1002          .addReg(DstReg,
1003                  RegState::Define |
1004                  getDeadRegState(isDead) |
1005                  getUndefRegState(isUndef), DstSubReg)
1006          .addFrameIndex(FI).addReg(0).addImm(0).addImm(Pred).addReg(PredReg);
1007      else // ARM::t2MOVr
1008        NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::t2LDRi12))
1009          .addReg(DstReg,
1010                  RegState::Define |
1011                  getDeadRegState(isDead) |
1012                  getUndefRegState(isUndef), DstSubReg)
1013          .addFrameIndex(FI).addImm(0).addImm(Pred).addReg(PredReg);
1014    }
1015  } else if (Opc == ARM::tMOVgpr2gpr ||
1016             Opc == ARM::tMOVtgpr2gpr ||
1017             Opc == ARM::tMOVgpr2tgpr) {
1018    if (OpNum == 0) { // move -> store
1019      unsigned SrcReg = MI->getOperand(1).getReg();
1020      unsigned SrcSubReg = MI->getOperand(1).getSubReg();
1021      bool isKill = MI->getOperand(1).isKill();
1022      bool isUndef = MI->getOperand(1).isUndef();
1023      NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::t2STRi12))
1024        .addReg(SrcReg,
1025                getKillRegState(isKill) | getUndefRegState(isUndef),
1026                SrcSubReg)
1027        .addFrameIndex(FI).addImm(0).addImm(ARMCC::AL).addReg(0);
1028    } else {          // move -> load
1029      unsigned DstReg = MI->getOperand(0).getReg();
1030      unsigned DstSubReg = MI->getOperand(0).getSubReg();
1031      bool isDead = MI->getOperand(0).isDead();
1032      bool isUndef = MI->getOperand(0).isUndef();
1033      NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::t2LDRi12))
1034        .addReg(DstReg,
1035                RegState::Define |
1036                getDeadRegState(isDead) |
1037                getUndefRegState(isUndef),
1038                DstSubReg)
1039        .addFrameIndex(FI).addImm(0).addImm(ARMCC::AL).addReg(0);
1040    }
1041  } else if (Opc == ARM::VMOVS) {
1042    unsigned Pred = MI->getOperand(2).getImm();
1043    unsigned PredReg = MI->getOperand(3).getReg();
1044    if (OpNum == 0) { // move -> store
1045      unsigned SrcReg = MI->getOperand(1).getReg();
1046      unsigned SrcSubReg = MI->getOperand(1).getSubReg();
1047      bool isKill = MI->getOperand(1).isKill();
1048      bool isUndef = MI->getOperand(1).isUndef();
1049      NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::VSTRS))
1050        .addReg(SrcReg, getKillRegState(isKill) | getUndefRegState(isUndef),
1051                SrcSubReg)
1052        .addFrameIndex(FI)
1053        .addImm(0).addImm(Pred).addReg(PredReg);
1054    } else {          // move -> load
1055      unsigned DstReg = MI->getOperand(0).getReg();
1056      unsigned DstSubReg = MI->getOperand(0).getSubReg();
1057      bool isDead = MI->getOperand(0).isDead();
1058      bool isUndef = MI->getOperand(0).isUndef();
1059      NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::VLDRS))
1060        .addReg(DstReg,
1061                RegState::Define |
1062                getDeadRegState(isDead) |
1063                getUndefRegState(isUndef),
1064                DstSubReg)
1065        .addFrameIndex(FI).addImm(0).addImm(Pred).addReg(PredReg);
1066    }
1067  } else if (Opc == ARM::VMOVD || Opc == ARM::VMOVDneon) {
1068    unsigned Pred = MI->getOperand(2).getImm();
1069    unsigned PredReg = MI->getOperand(3).getReg();
1070    if (OpNum == 0) { // move -> store
1071      unsigned SrcReg = MI->getOperand(1).getReg();
1072      unsigned SrcSubReg = MI->getOperand(1).getSubReg();
1073      bool isKill = MI->getOperand(1).isKill();
1074      bool isUndef = MI->getOperand(1).isUndef();
1075      NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::VSTRD))
1076        .addReg(SrcReg,
1077                getKillRegState(isKill) | getUndefRegState(isUndef),
1078                SrcSubReg)
1079        .addFrameIndex(FI).addImm(0).addImm(Pred).addReg(PredReg);
1080    } else {          // move -> load
1081      unsigned DstReg = MI->getOperand(0).getReg();
1082      unsigned DstSubReg = MI->getOperand(0).getSubReg();
1083      bool isDead = MI->getOperand(0).isDead();
1084      bool isUndef = MI->getOperand(0).isUndef();
1085      NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::VLDRD))
1086        .addReg(DstReg,
1087                RegState::Define |
1088                getDeadRegState(isDead) |
1089                getUndefRegState(isUndef),
1090                DstSubReg)
1091        .addFrameIndex(FI).addImm(0).addImm(Pred).addReg(PredReg);
1092    }
1093  }  else if (Opc == ARM::VMOVQ) {
1094    MachineFrameInfo &MFI = *MF.getFrameInfo();
1095    unsigned Pred = MI->getOperand(2).getImm();
1096    unsigned PredReg = MI->getOperand(3).getReg();
1097    if (OpNum == 0) { // move -> store
1098      unsigned SrcReg = MI->getOperand(1).getReg();
1099      unsigned SrcSubReg = MI->getOperand(1).getSubReg();
1100      bool isKill = MI->getOperand(1).isKill();
1101      bool isUndef = MI->getOperand(1).isUndef();
1102      if (MFI.getObjectAlignment(FI) >= 16 &&
1103          getRegisterInfo().canRealignStack(MF)) {
1104        NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::VST1q))
1105          .addFrameIndex(FI).addImm(128)
1106          .addReg(SrcReg,
1107                  getKillRegState(isKill) | getUndefRegState(isUndef),
1108                  SrcSubReg)
1109          .addImm(Pred).addReg(PredReg);
1110      } else {
1111        NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::VSTMQ))
1112          .addReg(SrcReg,
1113                  getKillRegState(isKill) | getUndefRegState(isUndef),
1114                  SrcSubReg)
1115          .addFrameIndex(FI).addImm(ARM_AM::getAM5Opc(ARM_AM::ia, 4))
1116          .addImm(Pred).addReg(PredReg);
1117      }
1118    } else {          // move -> load
1119      unsigned DstReg = MI->getOperand(0).getReg();
1120      unsigned DstSubReg = MI->getOperand(0).getSubReg();
1121      bool isDead = MI->getOperand(0).isDead();
1122      bool isUndef = MI->getOperand(0).isUndef();
1123      if (MFI.getObjectAlignment(FI) >= 16 &&
1124          getRegisterInfo().canRealignStack(MF)) {
1125        NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::VLD1q))
1126          .addReg(DstReg,
1127                  RegState::Define |
1128                  getDeadRegState(isDead) |
1129                  getUndefRegState(isUndef),
1130                  DstSubReg)
1131          .addFrameIndex(FI).addImm(128).addImm(Pred).addReg(PredReg);
1132      } else {
1133        NewMI = BuildMI(MF, MI->getDebugLoc(), get(ARM::VLDMQ))
1134          .addReg(DstReg,
1135                  RegState::Define |
1136                  getDeadRegState(isDead) |
1137                  getUndefRegState(isUndef),
1138                  DstSubReg)
1139          .addFrameIndex(FI).addImm(ARM_AM::getAM5Opc(ARM_AM::ia, 4))
1140          .addImm(Pred).addReg(PredReg);
1141      }
1142    }
1143  }
1144
1145  return NewMI;
1146}
1147
1148MachineInstr*
1149ARMBaseInstrInfo::foldMemoryOperandImpl(MachineFunction &MF,
1150                                        MachineInstr* MI,
1151                                        const SmallVectorImpl<unsigned> &Ops,
1152                                        MachineInstr* LoadMI) const {
1153  // FIXME
1154  return 0;
1155}
1156
1157bool
1158ARMBaseInstrInfo::canFoldMemoryOperand(const MachineInstr *MI,
1159                                   const SmallVectorImpl<unsigned> &Ops) const {
1160  if (Ops.size() != 1) return false;
1161
1162  unsigned Opc = MI->getOpcode();
1163  if (Opc == ARM::MOVr || Opc == ARM::t2MOVr) {
1164    // If it is updating CPSR, then it cannot be folded.
1165    return MI->getOperand(4).getReg() != ARM::CPSR ||
1166      MI->getOperand(4).isDead();
1167  } else if (Opc == ARM::tMOVgpr2gpr ||
1168             Opc == ARM::tMOVtgpr2gpr ||
1169             Opc == ARM::tMOVgpr2tgpr) {
1170    return true;
1171  } else if (Opc == ARM::VMOVS || Opc == ARM::VMOVD ||
1172             Opc == ARM::VMOVDneon || Opc == ARM::VMOVQ) {
1173    return true;
1174  }
1175
1176  // FIXME: VMOVQQ and VMOVQQQQ?
1177
1178  return false;
1179}
1180
1181/// Create a copy of a const pool value. Update CPI to the new index and return
1182/// the label UID.
1183static unsigned duplicateCPV(MachineFunction &MF, unsigned &CPI) {
1184  MachineConstantPool *MCP = MF.getConstantPool();
1185  ARMFunctionInfo *AFI = MF.getInfo<ARMFunctionInfo>();
1186
1187  const MachineConstantPoolEntry &MCPE = MCP->getConstants()[CPI];
1188  assert(MCPE.isMachineConstantPoolEntry() &&
1189         "Expecting a machine constantpool entry!");
1190  ARMConstantPoolValue *ACPV =
1191    static_cast<ARMConstantPoolValue*>(MCPE.Val.MachineCPVal);
1192
1193  unsigned PCLabelId = AFI->createConstPoolEntryUId();
1194  ARMConstantPoolValue *NewCPV = 0;
1195  if (ACPV->isGlobalValue())
1196    NewCPV = new ARMConstantPoolValue(ACPV->getGV(), PCLabelId,
1197                                      ARMCP::CPValue, 4);
1198  else if (ACPV->isExtSymbol())
1199    NewCPV = new ARMConstantPoolValue(MF.getFunction()->getContext(),
1200                                      ACPV->getSymbol(), PCLabelId, 4);
1201  else if (ACPV->isBlockAddress())
1202    NewCPV = new ARMConstantPoolValue(ACPV->getBlockAddress(), PCLabelId,
1203                                      ARMCP::CPBlockAddress, 4);
1204  else
1205    llvm_unreachable("Unexpected ARM constantpool value type!!");
1206  CPI = MCP->getConstantPoolIndex(NewCPV, MCPE.getAlignment());
1207  return PCLabelId;
1208}
1209
1210void ARMBaseInstrInfo::
1211reMaterialize(MachineBasicBlock &MBB,
1212              MachineBasicBlock::iterator I,
1213              unsigned DestReg, unsigned SubIdx,
1214              const MachineInstr *Orig,
1215              const TargetRegisterInfo &TRI) const {
1216  unsigned Opcode = Orig->getOpcode();
1217  switch (Opcode) {
1218  default: {
1219    MachineInstr *MI = MBB.getParent()->CloneMachineInstr(Orig);
1220    MI->substituteRegister(Orig->getOperand(0).getReg(), DestReg, SubIdx, TRI);
1221    MBB.insert(I, MI);
1222    break;
1223  }
1224  case ARM::tLDRpci_pic:
1225  case ARM::t2LDRpci_pic: {
1226    MachineFunction &MF = *MBB.getParent();
1227    unsigned CPI = Orig->getOperand(1).getIndex();
1228    unsigned PCLabelId = duplicateCPV(MF, CPI);
1229    MachineInstrBuilder MIB = BuildMI(MBB, I, Orig->getDebugLoc(), get(Opcode),
1230                                      DestReg)
1231      .addConstantPoolIndex(CPI).addImm(PCLabelId);
1232    (*MIB).setMemRefs(Orig->memoperands_begin(), Orig->memoperands_end());
1233    break;
1234  }
1235  }
1236}
1237
1238MachineInstr *
1239ARMBaseInstrInfo::duplicate(MachineInstr *Orig, MachineFunction &MF) const {
1240  MachineInstr *MI = TargetInstrInfoImpl::duplicate(Orig, MF);
1241  switch(Orig->getOpcode()) {
1242  case ARM::tLDRpci_pic:
1243  case ARM::t2LDRpci_pic: {
1244    unsigned CPI = Orig->getOperand(1).getIndex();
1245    unsigned PCLabelId = duplicateCPV(MF, CPI);
1246    Orig->getOperand(1).setIndex(CPI);
1247    Orig->getOperand(2).setImm(PCLabelId);
1248    break;
1249  }
1250  }
1251  return MI;
1252}
1253
1254bool ARMBaseInstrInfo::produceSameValue(const MachineInstr *MI0,
1255                                        const MachineInstr *MI1) const {
1256  int Opcode = MI0->getOpcode();
1257  if (Opcode == ARM::t2LDRpci ||
1258      Opcode == ARM::t2LDRpci_pic ||
1259      Opcode == ARM::tLDRpci ||
1260      Opcode == ARM::tLDRpci_pic) {
1261    if (MI1->getOpcode() != Opcode)
1262      return false;
1263    if (MI0->getNumOperands() != MI1->getNumOperands())
1264      return false;
1265
1266    const MachineOperand &MO0 = MI0->getOperand(1);
1267    const MachineOperand &MO1 = MI1->getOperand(1);
1268    if (MO0.getOffset() != MO1.getOffset())
1269      return false;
1270
1271    const MachineFunction *MF = MI0->getParent()->getParent();
1272    const MachineConstantPool *MCP = MF->getConstantPool();
1273    int CPI0 = MO0.getIndex();
1274    int CPI1 = MO1.getIndex();
1275    const MachineConstantPoolEntry &MCPE0 = MCP->getConstants()[CPI0];
1276    const MachineConstantPoolEntry &MCPE1 = MCP->getConstants()[CPI1];
1277    ARMConstantPoolValue *ACPV0 =
1278      static_cast<ARMConstantPoolValue*>(MCPE0.Val.MachineCPVal);
1279    ARMConstantPoolValue *ACPV1 =
1280      static_cast<ARMConstantPoolValue*>(MCPE1.Val.MachineCPVal);
1281    return ACPV0->hasSameValue(ACPV1);
1282  }
1283
1284  return MI0->isIdenticalTo(MI1, MachineInstr::IgnoreVRegDefs);
1285}
1286
1287/// getInstrPredicate - If instruction is predicated, returns its predicate
1288/// condition, otherwise returns AL. It also returns the condition code
1289/// register by reference.
1290ARMCC::CondCodes
1291llvm::getInstrPredicate(const MachineInstr *MI, unsigned &PredReg) {
1292  int PIdx = MI->findFirstPredOperandIdx();
1293  if (PIdx == -1) {
1294    PredReg = 0;
1295    return ARMCC::AL;
1296  }
1297
1298  PredReg = MI->getOperand(PIdx+1).getReg();
1299  return (ARMCC::CondCodes)MI->getOperand(PIdx).getImm();
1300}
1301
1302
1303int llvm::getMatchingCondBranchOpcode(int Opc) {
1304  if (Opc == ARM::B)
1305    return ARM::Bcc;
1306  else if (Opc == ARM::tB)
1307    return ARM::tBcc;
1308  else if (Opc == ARM::t2B)
1309      return ARM::t2Bcc;
1310
1311  llvm_unreachable("Unknown unconditional branch opcode!");
1312  return 0;
1313}
1314
1315
1316void llvm::emitARMRegPlusImmediate(MachineBasicBlock &MBB,
1317                               MachineBasicBlock::iterator &MBBI, DebugLoc dl,
1318                               unsigned DestReg, unsigned BaseReg, int NumBytes,
1319                               ARMCC::CondCodes Pred, unsigned PredReg,
1320                               const ARMBaseInstrInfo &TII) {
1321  bool isSub = NumBytes < 0;
1322  if (isSub) NumBytes = -NumBytes;
1323
1324  while (NumBytes) {
1325    unsigned RotAmt = ARM_AM::getSOImmValRotate(NumBytes);
1326    unsigned ThisVal = NumBytes & ARM_AM::rotr32(0xFF, RotAmt);
1327    assert(ThisVal && "Didn't extract field correctly");
1328
1329    // We will handle these bits from offset, clear them.
1330    NumBytes &= ~ThisVal;
1331
1332    assert(ARM_AM::getSOImmVal(ThisVal) != -1 && "Bit extraction didn't work?");
1333
1334    // Build the new ADD / SUB.
1335    unsigned Opc = isSub ? ARM::SUBri : ARM::ADDri;
1336    BuildMI(MBB, MBBI, dl, TII.get(Opc), DestReg)
1337      .addReg(BaseReg, RegState::Kill).addImm(ThisVal)
1338      .addImm((unsigned)Pred).addReg(PredReg).addReg(0);
1339    BaseReg = DestReg;
1340  }
1341}
1342
1343bool llvm::rewriteARMFrameIndex(MachineInstr &MI, unsigned FrameRegIdx,
1344                                unsigned FrameReg, int &Offset,
1345                                const ARMBaseInstrInfo &TII) {
1346  unsigned Opcode = MI.getOpcode();
1347  const TargetInstrDesc &Desc = MI.getDesc();
1348  unsigned AddrMode = (Desc.TSFlags & ARMII::AddrModeMask);
1349  bool isSub = false;
1350
1351  // Memory operands in inline assembly always use AddrMode2.
1352  if (Opcode == ARM::INLINEASM)
1353    AddrMode = ARMII::AddrMode2;
1354
1355  if (Opcode == ARM::ADDri) {
1356    Offset += MI.getOperand(FrameRegIdx+1).getImm();
1357    if (Offset == 0) {
1358      // Turn it into a move.
1359      MI.setDesc(TII.get(ARM::MOVr));
1360      MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg, false);
1361      MI.RemoveOperand(FrameRegIdx+1);
1362      Offset = 0;
1363      return true;
1364    } else if (Offset < 0) {
1365      Offset = -Offset;
1366      isSub = true;
1367      MI.setDesc(TII.get(ARM::SUBri));
1368    }
1369
1370    // Common case: small offset, fits into instruction.
1371    if (ARM_AM::getSOImmVal(Offset) != -1) {
1372      // Replace the FrameIndex with sp / fp
1373      MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg, false);
1374      MI.getOperand(FrameRegIdx+1).ChangeToImmediate(Offset);
1375      Offset = 0;
1376      return true;
1377    }
1378
1379    // Otherwise, pull as much of the immedidate into this ADDri/SUBri
1380    // as possible.
1381    unsigned RotAmt = ARM_AM::getSOImmValRotate(Offset);
1382    unsigned ThisImmVal = Offset & ARM_AM::rotr32(0xFF, RotAmt);
1383
1384    // We will handle these bits from offset, clear them.
1385    Offset &= ~ThisImmVal;
1386
1387    // Get the properly encoded SOImmVal field.
1388    assert(ARM_AM::getSOImmVal(ThisImmVal) != -1 &&
1389           "Bit extraction didn't work?");
1390    MI.getOperand(FrameRegIdx+1).ChangeToImmediate(ThisImmVal);
1391 } else {
1392    unsigned ImmIdx = 0;
1393    int InstrOffs = 0;
1394    unsigned NumBits = 0;
1395    unsigned Scale = 1;
1396    switch (AddrMode) {
1397    case ARMII::AddrMode2: {
1398      ImmIdx = FrameRegIdx+2;
1399      InstrOffs = ARM_AM::getAM2Offset(MI.getOperand(ImmIdx).getImm());
1400      if (ARM_AM::getAM2Op(MI.getOperand(ImmIdx).getImm()) == ARM_AM::sub)
1401        InstrOffs *= -1;
1402      NumBits = 12;
1403      break;
1404    }
1405    case ARMII::AddrMode3: {
1406      ImmIdx = FrameRegIdx+2;
1407      InstrOffs = ARM_AM::getAM3Offset(MI.getOperand(ImmIdx).getImm());
1408      if (ARM_AM::getAM3Op(MI.getOperand(ImmIdx).getImm()) == ARM_AM::sub)
1409        InstrOffs *= -1;
1410      NumBits = 8;
1411      break;
1412    }
1413    case ARMII::AddrMode4:
1414    case ARMII::AddrMode6:
1415      // Can't fold any offset even if it's zero.
1416      return false;
1417    case ARMII::AddrMode5: {
1418      ImmIdx = FrameRegIdx+1;
1419      InstrOffs = ARM_AM::getAM5Offset(MI.getOperand(ImmIdx).getImm());
1420      if (ARM_AM::getAM5Op(MI.getOperand(ImmIdx).getImm()) == ARM_AM::sub)
1421        InstrOffs *= -1;
1422      NumBits = 8;
1423      Scale = 4;
1424      break;
1425    }
1426    default:
1427      llvm_unreachable("Unsupported addressing mode!");
1428      break;
1429    }
1430
1431    Offset += InstrOffs * Scale;
1432    assert((Offset & (Scale-1)) == 0 && "Can't encode this offset!");
1433    if (Offset < 0) {
1434      Offset = -Offset;
1435      isSub = true;
1436    }
1437
1438    // Attempt to fold address comp. if opcode has offset bits
1439    if (NumBits > 0) {
1440      // Common case: small offset, fits into instruction.
1441      MachineOperand &ImmOp = MI.getOperand(ImmIdx);
1442      int ImmedOffset = Offset / Scale;
1443      unsigned Mask = (1 << NumBits) - 1;
1444      if ((unsigned)Offset <= Mask * Scale) {
1445        // Replace the FrameIndex with sp
1446        MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg, false);
1447        if (isSub)
1448          ImmedOffset |= 1 << NumBits;
1449        ImmOp.ChangeToImmediate(ImmedOffset);
1450        Offset = 0;
1451        return true;
1452      }
1453
1454      // Otherwise, it didn't fit. Pull in what we can to simplify the immed.
1455      ImmedOffset = ImmedOffset & Mask;
1456      if (isSub)
1457        ImmedOffset |= 1 << NumBits;
1458      ImmOp.ChangeToImmediate(ImmedOffset);
1459      Offset &= ~(Mask*Scale);
1460    }
1461  }
1462
1463  Offset = (isSub) ? -Offset : Offset;
1464  return Offset == 0;
1465}
1466