2014-05-24 20:50:23 +08:00
|
|
|
//===-- AArch6464FastISel.cpp - AArch64 FastISel implementation -----------===//
|
2014-03-29 18:18:08 +08:00
|
|
|
//
|
|
|
|
// The LLVM Compiler Infrastructure
|
|
|
|
//
|
|
|
|
// This file is distributed under the University of Illinois Open Source
|
|
|
|
// License. See LICENSE.TXT for details.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
2014-05-24 20:50:23 +08:00
|
|
|
// This file defines the AArch64-specific support for the FastISel class. Some
|
2014-03-29 18:18:08 +08:00
|
|
|
// of the target-specific code is generated by tablegen in the file
|
2014-05-24 20:50:23 +08:00
|
|
|
// AArch64GenFastISel.inc, which is #included here.
|
2014-03-29 18:18:08 +08:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
#include "AArch64.h"
|
|
|
|
#include "AArch64Subtarget.h"
|
2014-07-25 19:42:14 +08:00
|
|
|
#include "AArch64TargetMachine.h"
|
2014-05-24 20:50:23 +08:00
|
|
|
#include "MCTargetDesc/AArch64AddressingModes.h"
|
2014-08-02 02:39:24 +08:00
|
|
|
#include "llvm/Analysis/BranchProbabilityInfo.h"
|
2014-03-29 18:18:08 +08:00
|
|
|
#include "llvm/CodeGen/CallingConvLower.h"
|
|
|
|
#include "llvm/CodeGen/FastISel.h"
|
|
|
|
#include "llvm/CodeGen/FunctionLoweringInfo.h"
|
|
|
|
#include "llvm/CodeGen/MachineConstantPool.h"
|
|
|
|
#include "llvm/CodeGen/MachineFrameInfo.h"
|
|
|
|
#include "llvm/CodeGen/MachineInstrBuilder.h"
|
|
|
|
#include "llvm/CodeGen/MachineRegisterInfo.h"
|
|
|
|
#include "llvm/IR/CallingConv.h"
|
|
|
|
#include "llvm/IR/DataLayout.h"
|
|
|
|
#include "llvm/IR/DerivedTypes.h"
|
|
|
|
#include "llvm/IR/Function.h"
|
|
|
|
#include "llvm/IR/GetElementPtrTypeIterator.h"
|
|
|
|
#include "llvm/IR/GlobalAlias.h"
|
|
|
|
#include "llvm/IR/GlobalVariable.h"
|
|
|
|
#include "llvm/IR/Instructions.h"
|
|
|
|
#include "llvm/IR/IntrinsicInst.h"
|
|
|
|
#include "llvm/IR/Operator.h"
|
|
|
|
#include "llvm/Support/CommandLine.h"
|
|
|
|
using namespace llvm;
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
|
2014-09-16 06:33:11 +08:00
|
|
|
class AArch64FastISel final : public FastISel {
|
2014-03-29 18:18:08 +08:00
|
|
|
class Address {
|
|
|
|
public:
|
|
|
|
typedef enum {
|
|
|
|
RegBase,
|
|
|
|
FrameIndexBase
|
|
|
|
} BaseKind;
|
|
|
|
|
|
|
|
private:
|
|
|
|
BaseKind Kind;
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
AArch64_AM::ShiftExtendType ExtType;
|
2014-03-29 18:18:08 +08:00
|
|
|
union {
|
|
|
|
unsigned Reg;
|
|
|
|
int FI;
|
|
|
|
} Base;
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
unsigned OffsetReg;
|
|
|
|
unsigned Shift;
|
2014-03-29 18:18:08 +08:00
|
|
|
int64_t Offset;
|
2014-07-31 12:10:40 +08:00
|
|
|
const GlobalValue *GV;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
public:
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
Address() : Kind(RegBase), ExtType(AArch64_AM::InvalidShiftExtend),
|
|
|
|
OffsetReg(0), Shift(0), Offset(0), GV(nullptr) { Base.Reg = 0; }
|
2014-03-29 18:18:08 +08:00
|
|
|
void setKind(BaseKind K) { Kind = K; }
|
|
|
|
BaseKind getKind() const { return Kind; }
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
void setExtendType(AArch64_AM::ShiftExtendType E) { ExtType = E; }
|
|
|
|
AArch64_AM::ShiftExtendType getExtendType() const { return ExtType; }
|
2014-03-29 18:18:08 +08:00
|
|
|
bool isRegBase() const { return Kind == RegBase; }
|
|
|
|
bool isFIBase() const { return Kind == FrameIndexBase; }
|
|
|
|
void setReg(unsigned Reg) {
|
|
|
|
assert(isRegBase() && "Invalid base register access!");
|
|
|
|
Base.Reg = Reg;
|
|
|
|
}
|
|
|
|
unsigned getReg() const {
|
|
|
|
assert(isRegBase() && "Invalid base register access!");
|
|
|
|
return Base.Reg;
|
|
|
|
}
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
void setOffsetReg(unsigned Reg) {
|
|
|
|
assert(isRegBase() && "Invalid offset register access!");
|
|
|
|
OffsetReg = Reg;
|
|
|
|
}
|
|
|
|
unsigned getOffsetReg() const {
|
|
|
|
assert(isRegBase() && "Invalid offset register access!");
|
|
|
|
return OffsetReg;
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
void setFI(unsigned FI) {
|
|
|
|
assert(isFIBase() && "Invalid base frame index access!");
|
|
|
|
Base.FI = FI;
|
|
|
|
}
|
|
|
|
unsigned getFI() const {
|
|
|
|
assert(isFIBase() && "Invalid base frame index access!");
|
|
|
|
return Base.FI;
|
|
|
|
}
|
|
|
|
void setOffset(int64_t O) { Offset = O; }
|
|
|
|
int64_t getOffset() { return Offset; }
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
void setShift(unsigned S) { Shift = S; }
|
|
|
|
unsigned getShift() { return Shift; }
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-07-31 12:10:40 +08:00
|
|
|
void setGlobalValue(const GlobalValue *G) { GV = G; }
|
|
|
|
const GlobalValue *getGlobalValue() { return GV; }
|
2014-03-29 18:18:08 +08:00
|
|
|
};
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
/// Subtarget - Keep a pointer to the AArch64Subtarget around so that we can
|
2014-03-29 18:18:08 +08:00
|
|
|
/// make the right decision when generating code for different targets.
|
2014-05-24 20:50:23 +08:00
|
|
|
const AArch64Subtarget *Subtarget;
|
2014-03-29 18:18:08 +08:00
|
|
|
LLVMContext *Context;
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
bool fastLowerArguments() override;
|
|
|
|
bool fastLowerCall(CallLoweringInfo &CLI) override;
|
|
|
|
bool fastLowerIntrinsicCall(const IntrinsicInst *II) override;
|
2014-07-23 07:14:58 +08:00
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
private:
|
|
|
|
// Selection routines.
|
2014-09-03 09:38:36 +08:00
|
|
|
bool selectAddSub(const Instruction *I);
|
2014-09-16 05:27:56 +08:00
|
|
|
bool selectLogicalOp(const Instruction *I);
|
2014-09-16 07:20:17 +08:00
|
|
|
bool selectLoad(const Instruction *I);
|
|
|
|
bool selectStore(const Instruction *I);
|
|
|
|
bool selectBranch(const Instruction *I);
|
|
|
|
bool selectIndirectBr(const Instruction *I);
|
|
|
|
bool selectCmp(const Instruction *I);
|
|
|
|
bool selectSelect(const Instruction *I);
|
|
|
|
bool selectFPExt(const Instruction *I);
|
|
|
|
bool selectFPTrunc(const Instruction *I);
|
|
|
|
bool selectFPToInt(const Instruction *I, bool Signed);
|
|
|
|
bool selectIntToFP(const Instruction *I, bool Signed);
|
|
|
|
bool selectRem(const Instruction *I, unsigned ISDOpcode);
|
|
|
|
bool selectRet(const Instruction *I);
|
|
|
|
bool selectTrunc(const Instruction *I);
|
|
|
|
bool selectIntExt(const Instruction *I);
|
|
|
|
bool selectMul(const Instruction *I);
|
|
|
|
bool selectShift(const Instruction *I);
|
|
|
|
bool selectBitCast(const Instruction *I);
|
2014-09-16 06:07:49 +08:00
|
|
|
bool selectFRem(const Instruction *I);
|
2014-09-18 05:55:55 +08:00
|
|
|
bool selectSDiv(const Instruction *I);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Utility helper routines.
|
|
|
|
bool isTypeLegal(Type *Ty, MVT &VT);
|
2014-09-16 05:27:54 +08:00
|
|
|
bool isTypeSupported(Type *Ty, MVT &VT, bool IsVectorAllowed = false);
|
2014-08-29 08:19:21 +08:00
|
|
|
bool isValueAvailable(const Value *V) const;
|
2014-09-16 07:20:17 +08:00
|
|
|
bool computeAddress(const Value *Obj, Address &Addr, Type *Ty = nullptr);
|
|
|
|
bool computeCallAddress(const Value *V, Address &Addr);
|
|
|
|
bool simplifyAddress(Address &Addr, MVT VT);
|
|
|
|
void addLoadStoreOperands(Address &Addr, const MachineInstrBuilder &MIB,
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
unsigned Flags, unsigned ScaleFactor,
|
|
|
|
MachineMemOperand *MMO);
|
2014-09-16 07:20:17 +08:00
|
|
|
bool isMemCpySmall(uint64_t Len, unsigned Alignment);
|
|
|
|
bool tryEmitSmallMemCpy(Address Dest, Address Src, uint64_t Len,
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned Alignment);
|
2014-07-31 06:04:34 +08:00
|
|
|
bool foldXALUIntrinsic(AArch64CC::CondCode &CC, const Instruction *I,
|
|
|
|
const Value *Cond);
|
|
|
|
|
2014-08-20 06:29:55 +08:00
|
|
|
// Emit helper routines.
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned emitAddSub(bool UseAdd, MVT RetVT, const Value *LHS,
|
|
|
|
const Value *RHS, bool SetFlags = false,
|
|
|
|
bool WantResult = true, bool IsZExt = false);
|
|
|
|
unsigned emitAddSub_rr(bool UseAdd, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg, bool RHSIsKill,
|
|
|
|
bool SetFlags = false, bool WantResult = true);
|
|
|
|
unsigned emitAddSub_ri(bool UseAdd, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, uint64_t Imm, bool SetFlags = false,
|
|
|
|
bool WantResult = true);
|
2014-08-27 08:58:30 +08:00
|
|
|
unsigned emitAddSub_rs(bool UseAdd, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg, bool RHSIsKill,
|
|
|
|
AArch64_AM::ShiftExtendType ShiftType,
|
2014-09-03 09:38:36 +08:00
|
|
|
uint64_t ShiftImm, bool SetFlags = false,
|
|
|
|
bool WantResult = true);
|
2014-08-27 08:58:30 +08:00
|
|
|
unsigned emitAddSub_rx(bool UseAdd, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg, bool RHSIsKill,
|
|
|
|
AArch64_AM::ShiftExtendType ExtType,
|
2014-09-03 09:38:36 +08:00
|
|
|
uint64_t ShiftImm, bool SetFlags = false,
|
|
|
|
bool WantResult = true);
|
2014-08-20 06:29:55 +08:00
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Emit functions.
|
2014-08-20 06:29:55 +08:00
|
|
|
bool emitCmp(const Value *LHS, const Value *RHS, bool IsZExt);
|
|
|
|
bool emitICmp(MVT RetVT, const Value *LHS, const Value *RHS, bool IsZExt);
|
|
|
|
bool emitICmp_ri(MVT RetVT, unsigned LHSReg, bool LHSIsKill, uint64_t Imm);
|
|
|
|
bool emitFCmp(MVT RetVT, const Value *LHS, const Value *RHS);
|
2014-09-16 07:20:17 +08:00
|
|
|
bool emitLoad(MVT VT, unsigned &ResultReg, Address Addr,
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
MachineMemOperand *MMO = nullptr);
|
2014-09-16 07:20:17 +08:00
|
|
|
bool emitStore(MVT VT, unsigned SrcReg, Address Addr,
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
MachineMemOperand *MMO = nullptr);
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned emitIntExt(MVT SrcVT, unsigned SrcReg, MVT DestVT, bool isZExt);
|
|
|
|
unsigned emiti1Ext(unsigned SrcReg, MVT DestVT, bool isZExt);
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned emitAdd(MVT RetVT, const Value *LHS, const Value *RHS,
|
|
|
|
bool SetFlags = false, bool WantResult = true,
|
|
|
|
bool IsZExt = false);
|
|
|
|
unsigned emitSub(MVT RetVT, const Value *LHS, const Value *RHS,
|
|
|
|
bool SetFlags = false, bool WantResult = true,
|
|
|
|
bool IsZExt = false);
|
2014-08-20 06:29:55 +08:00
|
|
|
unsigned emitSubs_rr(MVT RetVT, unsigned LHSReg, bool LHSIsKill,
|
|
|
|
unsigned RHSReg, bool RHSIsKill, bool WantResult = true);
|
|
|
|
unsigned emitSubs_rs(MVT RetVT, unsigned LHSReg, bool LHSIsKill,
|
|
|
|
unsigned RHSReg, bool RHSIsKill,
|
|
|
|
AArch64_AM::ShiftExtendType ShiftType, uint64_t ShiftImm,
|
|
|
|
bool WantResult = true);
|
2014-09-04 09:29:18 +08:00
|
|
|
unsigned emitLogicalOp(unsigned ISDOpc, MVT RetVT, const Value *LHS,
|
|
|
|
const Value *RHS);
|
|
|
|
unsigned emitLogicalOp_ri(unsigned ISDOpc, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, uint64_t Imm);
|
|
|
|
unsigned emitLogicalOp_rs(unsigned ISDOpc, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg, bool RHSIsKill,
|
|
|
|
uint64_t ShiftImm);
|
|
|
|
unsigned emitAnd_ri(MVT RetVT, unsigned LHSReg, bool LHSIsKill, uint64_t Imm);
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned emitMul_rr(MVT RetVT, unsigned Op0, bool Op0IsKill,
|
|
|
|
unsigned Op1, bool Op1IsKill);
|
|
|
|
unsigned emitSMULL_rr(MVT RetVT, unsigned Op0, bool Op0IsKill,
|
|
|
|
unsigned Op1, bool Op1IsKill);
|
|
|
|
unsigned emitUMULL_rr(MVT RetVT, unsigned Op0, bool Op0IsKill,
|
|
|
|
unsigned Op1, bool Op1IsKill);
|
2014-08-22 07:06:07 +08:00
|
|
|
unsigned emitLSL_rr(MVT RetVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
unsigned Op1Reg, bool Op1IsKill);
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned emitLSL_ri(MVT RetVT, MVT SrcVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
uint64_t Imm, bool IsZExt = true);
|
2014-08-22 07:06:07 +08:00
|
|
|
unsigned emitLSR_rr(MVT RetVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
unsigned Op1Reg, bool Op1IsKill);
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned emitLSR_ri(MVT RetVT, MVT SrcVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
uint64_t Imm, bool IsZExt = true);
|
2014-08-22 07:06:07 +08:00
|
|
|
unsigned emitASR_rr(MVT RetVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
unsigned Op1Reg, bool Op1IsKill);
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned emitASR_ri(MVT RetVT, MVT SrcVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
uint64_t Imm, bool IsZExt = false);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned materializeInt(const ConstantInt *CI, MVT VT);
|
|
|
|
unsigned materializeFP(const ConstantFP *CFP, MVT VT);
|
|
|
|
unsigned materializeGV(const GlobalValue *GV);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Call handling routines.
|
|
|
|
private:
|
|
|
|
CCAssignFn *CCAssignFnForCall(CallingConv::ID CC) const;
|
2014-09-16 07:20:17 +08:00
|
|
|
bool processCallArgs(CallLoweringInfo &CLI, SmallVectorImpl<MVT> &ArgVTs,
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned &NumBytes);
|
2014-09-16 07:20:17 +08:00
|
|
|
bool finishCall(CallLoweringInfo &CLI, MVT RetVT, unsigned NumBytes);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
public:
|
|
|
|
// Backend specific FastISel code.
|
2014-09-04 04:56:52 +08:00
|
|
|
unsigned fastMaterializeAlloca(const AllocaInst *AI) override;
|
|
|
|
unsigned fastMaterializeConstant(const Constant *C) override;
|
|
|
|
unsigned fastMaterializeFloatZero(const ConstantFP* CF) override;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-09-03 05:32:54 +08:00
|
|
|
explicit AArch64FastISel(FunctionLoweringInfo &FuncInfo,
|
|
|
|
const TargetLibraryInfo *LibInfo)
|
|
|
|
: FastISel(FuncInfo, LibInfo, /*SkipTargetIndependentISel=*/true) {
|
2014-05-24 20:50:23 +08:00
|
|
|
Subtarget = &TM.getSubtarget<AArch64Subtarget>();
|
2014-09-03 05:32:54 +08:00
|
|
|
Context = &FuncInfo.Fn->getContext();
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
bool fastSelectInstruction(const Instruction *I) override;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
#include "AArch64GenFastISel.inc"
|
2014-03-29 18:18:08 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
} // end anonymous namespace
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
#include "AArch64GenCallingConv.inc"
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-09-30 08:49:54 +08:00
|
|
|
/// \brief Determine the implicit scale factor that is applied by a memory
|
|
|
|
/// operation for a given value type.
|
|
|
|
static unsigned getImplicitScaleFactor(MVT VT) {
|
|
|
|
switch (VT.SimpleTy) {
|
|
|
|
default:
|
|
|
|
return 0; // invalid
|
|
|
|
case MVT::i1: // fall-through
|
|
|
|
case MVT::i8:
|
|
|
|
return 1;
|
|
|
|
case MVT::i16:
|
|
|
|
return 2;
|
|
|
|
case MVT::i32: // fall-through
|
|
|
|
case MVT::f32:
|
|
|
|
return 4;
|
|
|
|
case MVT::i64: // fall-through
|
|
|
|
case MVT::f64:
|
|
|
|
return 8;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
CCAssignFn *AArch64FastISel::CCAssignFnForCall(CallingConv::ID CC) const {
|
2014-03-29 18:18:08 +08:00
|
|
|
if (CC == CallingConv::WebKit_JS)
|
2014-05-24 20:50:23 +08:00
|
|
|
return CC_AArch64_WebKit_JS;
|
|
|
|
return Subtarget->isTargetDarwin() ? CC_AArch64_DarwinPCS : CC_AArch64_AAPCS;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
unsigned AArch64FastISel::fastMaterializeAlloca(const AllocaInst *AI) {
|
2014-03-29 18:18:08 +08:00
|
|
|
assert(TLI.getValueType(AI->getType(), true) == MVT::i64 &&
|
|
|
|
"Alloca should always return a pointer.");
|
|
|
|
|
|
|
|
// Don't handle dynamic allocas.
|
|
|
|
if (!FuncInfo.StaticAllocaMap.count(AI))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
DenseMap<const AllocaInst *, int>::iterator SI =
|
|
|
|
FuncInfo.StaticAllocaMap.find(AI);
|
|
|
|
|
|
|
|
if (SI != FuncInfo.StaticAllocaMap.end()) {
|
2014-08-22 04:57:57 +08:00
|
|
|
unsigned ResultReg = createResultReg(&AArch64::GPR64spRegClass);
|
2014-05-24 20:50:23 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::ADDXri),
|
2014-03-29 18:18:08 +08:00
|
|
|
ResultReg)
|
|
|
|
.addFrameIndex(SI->second)
|
|
|
|
.addImm(0)
|
|
|
|
.addImm(0);
|
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned AArch64FastISel::materializeInt(const ConstantInt *CI, MVT VT) {
|
2014-08-16 02:55:52 +08:00
|
|
|
if (VT > MVT::i64)
|
|
|
|
return 0;
|
2014-08-20 03:44:02 +08:00
|
|
|
|
|
|
|
if (!CI->isZero())
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmit_i(VT, VT, ISD::Constant, CI->getZExtValue());
|
2014-08-20 03:44:02 +08:00
|
|
|
|
|
|
|
// Create a copy from the zero register to materialize a "0" value.
|
|
|
|
const TargetRegisterClass *RC = (VT == MVT::i64) ? &AArch64::GPR64RegClass
|
|
|
|
: &AArch64::GPR32RegClass;
|
|
|
|
unsigned ZeroReg = (VT == MVT::i64) ? AArch64::XZR : AArch64::WZR;
|
|
|
|
unsigned ResultReg = createResultReg(RC);
|
2014-08-22 04:57:57 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(TargetOpcode::COPY),
|
|
|
|
ResultReg).addReg(ZeroReg, getKillRegState(true));
|
2014-08-20 03:44:02 +08:00
|
|
|
return ResultReg;
|
2014-08-16 02:55:52 +08:00
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned AArch64FastISel::materializeFP(const ConstantFP *CFP, MVT VT) {
|
2014-08-26 03:58:05 +08:00
|
|
|
// Positive zero (+0.0) has to be materialized with a fmov from the zero
|
|
|
|
// register, because the immediate version of fmov cannot encode zero.
|
|
|
|
if (CFP->isNullValue())
|
2014-09-04 04:56:52 +08:00
|
|
|
return fastMaterializeFloatZero(CFP);
|
2014-08-26 03:58:05 +08:00
|
|
|
|
2014-04-30 23:29:57 +08:00
|
|
|
if (VT != MVT::f32 && VT != MVT::f64)
|
|
|
|
return 0;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
const APFloat Val = CFP->getValueAPF();
|
2014-08-16 02:55:52 +08:00
|
|
|
bool Is64Bit = (VT == MVT::f64);
|
2014-03-29 18:18:08 +08:00
|
|
|
// This checks to see if we can use FMOV instructions to materialize
|
|
|
|
// a constant, otherwise we have to materialize via the constant pool.
|
|
|
|
if (TLI.isFPImmLegal(Val, VT)) {
|
2014-08-26 03:58:05 +08:00
|
|
|
int Imm =
|
|
|
|
Is64Bit ? AArch64_AM::getFP64Imm(Val) : AArch64_AM::getFP32Imm(Val);
|
|
|
|
assert((Imm != -1) && "Cannot encode floating-point constant.");
|
2014-08-16 02:55:52 +08:00
|
|
|
unsigned Opc = Is64Bit ? AArch64::FMOVDi : AArch64::FMOVSi;
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_i(Opc, TLI.getRegClassFor(VT), Imm);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Materialize via constant pool. MachineConstantPool wants an explicit
|
|
|
|
// alignment.
|
|
|
|
unsigned Align = DL.getPrefTypeAlignment(CFP->getType());
|
|
|
|
if (Align == 0)
|
|
|
|
Align = DL.getTypeAllocSize(CFP->getType());
|
|
|
|
|
2014-08-16 02:55:52 +08:00
|
|
|
unsigned CPI = MCP.getConstantPoolIndex(cast<Constant>(CFP), Align);
|
2014-05-24 20:50:23 +08:00
|
|
|
unsigned ADRPReg = createResultReg(&AArch64::GPR64commonRegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::ADRP),
|
2014-08-26 03:58:05 +08:00
|
|
|
ADRPReg).addConstantPoolIndex(CPI, 0, AArch64II::MO_PAGE);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-08-16 02:55:52 +08:00
|
|
|
unsigned Opc = Is64Bit ? AArch64::LDRDui : AArch64::LDRSui;
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned ResultReg = createResultReg(TLI.getRegClassFor(VT));
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(Opc), ResultReg)
|
2014-08-26 03:58:05 +08:00
|
|
|
.addReg(ADRPReg)
|
|
|
|
.addConstantPoolIndex(CPI, 0, AArch64II::MO_PAGEOFF | AArch64II::MO_NC);
|
2014-03-29 18:18:08 +08:00
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned AArch64FastISel::materializeGV(const GlobalValue *GV) {
|
2014-05-29 02:15:43 +08:00
|
|
|
// We can't handle thread-local variables quickly yet.
|
|
|
|
if (GV->isThreadLocal())
|
|
|
|
return 0;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-05-25 03:45:41 +08:00
|
|
|
// MachO still uses GOT for large code-model accesses, but ELF requires
|
|
|
|
// movz/movk sequences, which FastISel doesn't handle yet.
|
|
|
|
if (TM.getCodeModel() != CodeModel::Small && !Subtarget->isTargetMachO())
|
|
|
|
return 0;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned char OpFlags = Subtarget->ClassifyGlobalReference(GV, TM);
|
|
|
|
|
|
|
|
EVT DestEVT = TLI.getValueType(GV->getType(), true);
|
|
|
|
if (!DestEVT.isSimple())
|
|
|
|
return 0;
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
unsigned ADRPReg = createResultReg(&AArch64::GPR64commonRegClass);
|
2014-04-15 21:59:53 +08:00
|
|
|
unsigned ResultReg;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
if (OpFlags & AArch64II::MO_GOT) {
|
2014-03-29 18:18:08 +08:00
|
|
|
// ADRP + LDRX
|
2014-05-24 20:50:23 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::ADRP),
|
2014-03-29 18:18:08 +08:00
|
|
|
ADRPReg)
|
2014-08-16 02:55:52 +08:00
|
|
|
.addGlobalAddress(GV, 0, AArch64II::MO_GOT | AArch64II::MO_PAGE);
|
2014-04-15 21:59:53 +08:00
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
ResultReg = createResultReg(&AArch64::GPR64RegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::LDRXui),
|
2014-03-29 18:18:08 +08:00
|
|
|
ResultReg)
|
2014-08-16 02:55:52 +08:00
|
|
|
.addReg(ADRPReg)
|
|
|
|
.addGlobalAddress(GV, 0, AArch64II::MO_GOT | AArch64II::MO_PAGEOFF |
|
|
|
|
AArch64II::MO_NC);
|
2014-09-10 21:54:38 +08:00
|
|
|
} else if (OpFlags & AArch64II::MO_CONSTPOOL) {
|
|
|
|
// We can't handle addresses loaded from a constant pool quickly yet.
|
|
|
|
return 0;
|
2014-03-29 18:18:08 +08:00
|
|
|
} else {
|
|
|
|
// ADRP + ADDX
|
2014-05-24 20:50:23 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::ADRP),
|
2014-08-16 02:55:52 +08:00
|
|
|
ADRPReg)
|
|
|
|
.addGlobalAddress(GV, 0, AArch64II::MO_PAGE);
|
2014-04-15 21:59:53 +08:00
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
ResultReg = createResultReg(&AArch64::GPR64spRegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::ADDXri),
|
2014-03-29 18:18:08 +08:00
|
|
|
ResultReg)
|
2014-08-16 02:55:52 +08:00
|
|
|
.addReg(ADRPReg)
|
|
|
|
.addGlobalAddress(GV, 0, AArch64II::MO_PAGEOFF | AArch64II::MO_NC)
|
|
|
|
.addImm(0);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
unsigned AArch64FastISel::fastMaterializeConstant(const Constant *C) {
|
2014-03-29 18:18:08 +08:00
|
|
|
EVT CEVT = TLI.getValueType(C->getType(), true);
|
|
|
|
|
|
|
|
// Only handle simple types.
|
|
|
|
if (!CEVT.isSimple())
|
|
|
|
return 0;
|
|
|
|
MVT VT = CEVT.getSimpleVT();
|
|
|
|
|
2014-08-16 02:55:52 +08:00
|
|
|
if (const auto *CI = dyn_cast<ConstantInt>(C))
|
2014-09-16 07:20:17 +08:00
|
|
|
return materializeInt(CI, VT);
|
2014-08-16 02:55:52 +08:00
|
|
|
else if (const ConstantFP *CFP = dyn_cast<ConstantFP>(C))
|
2014-09-16 07:20:17 +08:00
|
|
|
return materializeFP(CFP, VT);
|
2014-03-29 18:18:08 +08:00
|
|
|
else if (const GlobalValue *GV = dyn_cast<GlobalValue>(C))
|
2014-09-16 07:20:17 +08:00
|
|
|
return materializeGV(GV);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
unsigned AArch64FastISel::fastMaterializeFloatZero(const ConstantFP* CFP) {
|
2014-08-26 03:58:05 +08:00
|
|
|
assert(CFP->isNullValue() &&
|
|
|
|
"Floating-point constant is not a positive zero.");
|
|
|
|
MVT VT;
|
|
|
|
if (!isTypeLegal(CFP->getType(), VT))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (VT != MVT::f32 && VT != MVT::f64)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
bool Is64Bit = (VT == MVT::f64);
|
|
|
|
unsigned ZReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
|
|
|
|
unsigned Opc = Is64Bit ? AArch64::FMOVXDr : AArch64::FMOVWSr;
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_r(Opc, TLI.getRegClassFor(VT), ZReg, /*IsKill=*/true);
|
2014-08-26 03:58:05 +08:00
|
|
|
}
|
|
|
|
|
2014-09-18 03:19:31 +08:00
|
|
|
/// \brief Check if the multiply is by a power-of-2 constant.
|
|
|
|
static bool isMulPowOf2(const Value *I) {
|
|
|
|
if (const auto *MI = dyn_cast<MulOperator>(I)) {
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(MI->getOperand(0)))
|
|
|
|
if (C->getValue().isPowerOf2())
|
|
|
|
return true;
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(MI->getOperand(1)))
|
|
|
|
if (C->getValue().isPowerOf2())
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Computes the address to get to an object.
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::computeAddress(const Value *Obj, Address &Addr, Type *Ty)
|
2014-08-28 07:09:40 +08:00
|
|
|
{
|
2014-04-25 13:30:21 +08:00
|
|
|
const User *U = nullptr;
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned Opcode = Instruction::UserOp1;
|
|
|
|
if (const Instruction *I = dyn_cast<Instruction>(Obj)) {
|
|
|
|
// Don't walk into other basic blocks unless the object is an alloca from
|
|
|
|
// another block, otherwise it may not have a virtual register assigned.
|
|
|
|
if (FuncInfo.StaticAllocaMap.count(static_cast<const AllocaInst *>(Obj)) ||
|
|
|
|
FuncInfo.MBBMap[I->getParent()] == FuncInfo.MBB) {
|
|
|
|
Opcode = I->getOpcode();
|
|
|
|
U = I;
|
|
|
|
}
|
|
|
|
} else if (const ConstantExpr *C = dyn_cast<ConstantExpr>(Obj)) {
|
|
|
|
Opcode = C->getOpcode();
|
|
|
|
U = C;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (const PointerType *Ty = dyn_cast<PointerType>(Obj->getType()))
|
|
|
|
if (Ty->getAddressSpace() > 255)
|
|
|
|
// Fast instruction selection doesn't support the special
|
|
|
|
// address spaces.
|
|
|
|
return false;
|
|
|
|
|
|
|
|
switch (Opcode) {
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
case Instruction::BitCast: {
|
|
|
|
// Look through bitcasts.
|
2014-09-16 07:20:17 +08:00
|
|
|
return computeAddress(U->getOperand(0), Addr, Ty);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
case Instruction::IntToPtr: {
|
|
|
|
// Look past no-op inttoptrs.
|
|
|
|
if (TLI.getValueType(U->getOperand(0)->getType()) == TLI.getPointerTy())
|
2014-09-16 07:20:17 +08:00
|
|
|
return computeAddress(U->getOperand(0), Addr, Ty);
|
2014-03-29 18:18:08 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
case Instruction::PtrToInt: {
|
2014-08-28 07:09:40 +08:00
|
|
|
// Look past no-op ptrtoints.
|
2014-03-29 18:18:08 +08:00
|
|
|
if (TLI.getValueType(U->getType()) == TLI.getPointerTy())
|
2014-09-16 07:20:17 +08:00
|
|
|
return computeAddress(U->getOperand(0), Addr, Ty);
|
2014-03-29 18:18:08 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
case Instruction::GetElementPtr: {
|
|
|
|
Address SavedAddr = Addr;
|
|
|
|
uint64_t TmpOffset = Addr.getOffset();
|
|
|
|
|
|
|
|
// Iterate through the GEP folding the constants into offsets where
|
|
|
|
// we can.
|
|
|
|
gep_type_iterator GTI = gep_type_begin(U);
|
|
|
|
for (User::const_op_iterator i = U->op_begin() + 1, e = U->op_end(); i != e;
|
|
|
|
++i, ++GTI) {
|
|
|
|
const Value *Op = *i;
|
|
|
|
if (StructType *STy = dyn_cast<StructType>(*GTI)) {
|
|
|
|
const StructLayout *SL = DL.getStructLayout(STy);
|
|
|
|
unsigned Idx = cast<ConstantInt>(Op)->getZExtValue();
|
|
|
|
TmpOffset += SL->getElementOffset(Idx);
|
|
|
|
} else {
|
|
|
|
uint64_t S = DL.getTypeAllocSize(GTI.getIndexedType());
|
|
|
|
for (;;) {
|
|
|
|
if (const ConstantInt *CI = dyn_cast<ConstantInt>(Op)) {
|
|
|
|
// Constant-offset addressing.
|
|
|
|
TmpOffset += CI->getSExtValue() * S;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (canFoldAddIntoGEP(U, Op)) {
|
|
|
|
// A compatible add with a constant operand. Fold the constant.
|
|
|
|
ConstantInt *CI =
|
|
|
|
cast<ConstantInt>(cast<AddOperator>(Op)->getOperand(1));
|
|
|
|
TmpOffset += CI->getSExtValue() * S;
|
|
|
|
// Iterate on the other operand.
|
|
|
|
Op = cast<AddOperator>(Op)->getOperand(0);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
// Unsupported
|
|
|
|
goto unsupported_gep;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Try to grab the base operand now.
|
|
|
|
Addr.setOffset(TmpOffset);
|
2014-09-16 07:20:17 +08:00
|
|
|
if (computeAddress(U->getOperand(0), Addr, Ty))
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
|
|
|
|
// We failed, restore everything and try the other options.
|
|
|
|
Addr = SavedAddr;
|
|
|
|
|
|
|
|
unsupported_gep:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case Instruction::Alloca: {
|
|
|
|
const AllocaInst *AI = cast<AllocaInst>(Obj);
|
|
|
|
DenseMap<const AllocaInst *, int>::iterator SI =
|
|
|
|
FuncInfo.StaticAllocaMap.find(AI);
|
|
|
|
if (SI != FuncInfo.StaticAllocaMap.end()) {
|
|
|
|
Addr.setKind(Address::FrameIndexBase);
|
|
|
|
Addr.setFI(SI->second);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
case Instruction::Add: {
|
2014-08-02 03:40:16 +08:00
|
|
|
// Adds of constants are common and easy enough.
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
const Value *LHS = U->getOperand(0);
|
|
|
|
const Value *RHS = U->getOperand(1);
|
|
|
|
|
|
|
|
if (isa<ConstantInt>(LHS))
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
|
|
|
if (const ConstantInt *CI = dyn_cast<ConstantInt>(RHS)) {
|
2014-08-02 03:40:16 +08:00
|
|
|
Addr.setOffset(Addr.getOffset() + (uint64_t)CI->getSExtValue());
|
2014-09-16 07:20:17 +08:00
|
|
|
return computeAddress(LHS, Addr, Ty);
|
2014-08-02 03:40:16 +08:00
|
|
|
}
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
|
|
|
|
Address Backup = Addr;
|
2014-09-16 07:20:17 +08:00
|
|
|
if (computeAddress(LHS, Addr, Ty) && computeAddress(RHS, Addr, Ty))
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
return true;
|
|
|
|
Addr = Backup;
|
|
|
|
|
2014-08-02 03:40:16 +08:00
|
|
|
break;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
2014-09-20 06:23:46 +08:00
|
|
|
case Instruction::Shl: {
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
if (Addr.getOffsetReg())
|
|
|
|
break;
|
|
|
|
|
|
|
|
if (const auto *CI = dyn_cast<ConstantInt>(U->getOperand(1))) {
|
|
|
|
unsigned Val = CI->getZExtValue();
|
|
|
|
if (Val < 1 || Val > 3)
|
|
|
|
break;
|
|
|
|
|
|
|
|
uint64_t NumBytes = 0;
|
|
|
|
if (Ty && Ty->isSized()) {
|
|
|
|
uint64_t NumBits = DL.getTypeSizeInBits(Ty);
|
|
|
|
NumBytes = NumBits / 8;
|
|
|
|
if (!isPowerOf2_64(NumBits))
|
|
|
|
NumBytes = 0;
|
|
|
|
}
|
|
|
|
|
2014-08-20 20:14:35 +08:00
|
|
|
if (NumBytes != (1ULL << Val))
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
break;
|
|
|
|
|
|
|
|
Addr.setShift(Val);
|
|
|
|
Addr.setExtendType(AArch64_AM::LSL);
|
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
const Value *Src = U->getOperand(0);
|
|
|
|
if (const auto *I = dyn_cast<Instruction>(Src))
|
2014-08-28 07:09:40 +08:00
|
|
|
if (FuncInfo.MBBMap[I->getParent()] == FuncInfo.MBB)
|
2014-09-20 06:23:46 +08:00
|
|
|
Src = I;
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
if (const auto *ZE = dyn_cast<ZExtInst>(Src)) {
|
|
|
|
if (ZE->getOperand(0)->getType()->isIntegerTy(32)) {
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
Addr.setExtendType(AArch64_AM::UXTW);
|
2014-09-20 06:23:46 +08:00
|
|
|
Src = ZE->getOperand(0);
|
|
|
|
}
|
|
|
|
} else if (const auto *SE = dyn_cast<SExtInst>(Src)) {
|
|
|
|
if (SE->getOperand(0)->getType()->isIntegerTy(32)) {
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
Addr.setExtendType(AArch64_AM::SXTW);
|
2014-09-20 06:23:46 +08:00
|
|
|
Src = SE->getOperand(0);
|
|
|
|
}
|
|
|
|
}
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
if (const auto *AI = dyn_cast<BinaryOperator>(Src))
|
2014-09-18 13:40:41 +08:00
|
|
|
if (AI->getOpcode() == Instruction::And) {
|
|
|
|
const Value *LHS = AI->getOperand(0);
|
|
|
|
const Value *RHS = AI->getOperand(1);
|
|
|
|
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(LHS))
|
|
|
|
if (C->getValue() == 0xffffffff)
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(RHS))
|
2014-09-18 13:40:41 +08:00
|
|
|
if (C->getValue() == 0xffffffff) {
|
|
|
|
Addr.setExtendType(AArch64_AM::UXTW);
|
|
|
|
unsigned Reg = getRegForValue(LHS);
|
|
|
|
if (!Reg)
|
|
|
|
return false;
|
|
|
|
bool RegIsKill = hasTrivialKill(LHS);
|
|
|
|
Reg = fastEmitInst_extractsubreg(MVT::i32, Reg, RegIsKill,
|
|
|
|
AArch64::sub_32);
|
|
|
|
Addr.setOffsetReg(Reg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
unsigned Reg = getRegForValue(Src);
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
if (!Reg)
|
|
|
|
return false;
|
|
|
|
Addr.setOffsetReg(Reg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
break;
|
2014-09-20 06:23:46 +08:00
|
|
|
}
|
2014-09-18 03:19:31 +08:00
|
|
|
case Instruction::Mul: {
|
|
|
|
if (Addr.getOffsetReg())
|
|
|
|
break;
|
|
|
|
|
|
|
|
if (!isMulPowOf2(U))
|
|
|
|
break;
|
|
|
|
|
|
|
|
const Value *LHS = U->getOperand(0);
|
|
|
|
const Value *RHS = U->getOperand(1);
|
|
|
|
|
|
|
|
// Canonicalize power-of-2 value to the RHS.
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(LHS))
|
|
|
|
if (C->getValue().isPowerOf2())
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
|
|
|
assert(isa<ConstantInt>(RHS) && "Expected an ConstantInt.");
|
|
|
|
const auto *C = cast<ConstantInt>(RHS);
|
|
|
|
unsigned Val = C->getValue().logBase2();
|
|
|
|
if (Val < 1 || Val > 3)
|
|
|
|
break;
|
|
|
|
|
|
|
|
uint64_t NumBytes = 0;
|
|
|
|
if (Ty && Ty->isSized()) {
|
|
|
|
uint64_t NumBits = DL.getTypeSizeInBits(Ty);
|
|
|
|
NumBytes = NumBits / 8;
|
|
|
|
if (!isPowerOf2_64(NumBits))
|
|
|
|
NumBytes = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (NumBytes != (1ULL << Val))
|
|
|
|
break;
|
|
|
|
|
|
|
|
Addr.setShift(Val);
|
|
|
|
Addr.setExtendType(AArch64_AM::LSL);
|
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
const Value *Src = LHS;
|
|
|
|
if (const auto *I = dyn_cast<Instruction>(Src))
|
2014-09-18 03:19:31 +08:00
|
|
|
if (FuncInfo.MBBMap[I->getParent()] == FuncInfo.MBB)
|
2014-09-20 06:23:46 +08:00
|
|
|
Src = I;
|
2014-09-18 03:19:31 +08:00
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
if (const auto *ZE = dyn_cast<ZExtInst>(Src)) {
|
2014-09-18 03:19:31 +08:00
|
|
|
if (ZE->getOperand(0)->getType()->isIntegerTy(32)) {
|
|
|
|
Addr.setExtendType(AArch64_AM::UXTW);
|
2014-09-20 06:23:46 +08:00
|
|
|
Src = ZE->getOperand(0);
|
2014-09-18 03:19:31 +08:00
|
|
|
}
|
2014-09-20 06:23:46 +08:00
|
|
|
} else if (const auto *SE = dyn_cast<SExtInst>(Src)) {
|
2014-09-18 03:19:31 +08:00
|
|
|
if (SE->getOperand(0)->getType()->isIntegerTy(32)) {
|
|
|
|
Addr.setExtendType(AArch64_AM::SXTW);
|
2014-09-20 06:23:46 +08:00
|
|
|
Src = SE->getOperand(0);
|
2014-09-18 03:19:31 +08:00
|
|
|
}
|
2014-09-20 06:23:46 +08:00
|
|
|
}
|
2014-09-18 03:19:31 +08:00
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
unsigned Reg = getRegForValue(Src);
|
2014-09-18 03:19:31 +08:00
|
|
|
if (!Reg)
|
|
|
|
return false;
|
|
|
|
Addr.setOffsetReg(Reg);
|
|
|
|
return true;
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
}
|
2014-09-18 13:40:41 +08:00
|
|
|
case Instruction::And: {
|
|
|
|
if (Addr.getOffsetReg())
|
|
|
|
break;
|
|
|
|
|
|
|
|
if (DL.getTypeSizeInBits(Ty) != 8)
|
|
|
|
break;
|
|
|
|
|
|
|
|
const Value *LHS = U->getOperand(0);
|
|
|
|
const Value *RHS = U->getOperand(1);
|
|
|
|
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(LHS))
|
|
|
|
if (C->getValue() == 0xffffffff)
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
2014-09-20 06:23:46 +08:00
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(RHS))
|
2014-09-18 13:40:41 +08:00
|
|
|
if (C->getValue() == 0xffffffff) {
|
|
|
|
Addr.setShift(0);
|
|
|
|
Addr.setExtendType(AArch64_AM::LSL);
|
|
|
|
Addr.setExtendType(AArch64_AM::UXTW);
|
|
|
|
|
|
|
|
unsigned Reg = getRegForValue(LHS);
|
|
|
|
if (!Reg)
|
|
|
|
return false;
|
|
|
|
bool RegIsKill = hasTrivialKill(LHS);
|
|
|
|
Reg = fastEmitInst_extractsubreg(MVT::i32, Reg, RegIsKill,
|
|
|
|
AArch64::sub_32);
|
|
|
|
Addr.setOffsetReg(Reg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
2014-09-18 03:19:31 +08:00
|
|
|
} // end switch
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
|
|
|
|
if (Addr.getReg()) {
|
|
|
|
if (!Addr.getOffsetReg()) {
|
|
|
|
unsigned Reg = getRegForValue(Obj);
|
|
|
|
if (!Reg)
|
|
|
|
return false;
|
|
|
|
Addr.setOffsetReg(Reg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
unsigned Reg = getRegForValue(Obj);
|
|
|
|
if (!Reg)
|
|
|
|
return false;
|
|
|
|
Addr.setReg(Reg);
|
|
|
|
return true;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::computeCallAddress(const Value *V, Address &Addr) {
|
2014-07-31 12:10:40 +08:00
|
|
|
const User *U = nullptr;
|
|
|
|
unsigned Opcode = Instruction::UserOp1;
|
|
|
|
bool InMBB = true;
|
|
|
|
|
|
|
|
if (const auto *I = dyn_cast<Instruction>(V)) {
|
|
|
|
Opcode = I->getOpcode();
|
|
|
|
U = I;
|
|
|
|
InMBB = I->getParent() == FuncInfo.MBB->getBasicBlock();
|
|
|
|
} else if (const auto *C = dyn_cast<ConstantExpr>(V)) {
|
|
|
|
Opcode = C->getOpcode();
|
|
|
|
U = C;
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (Opcode) {
|
|
|
|
default: break;
|
|
|
|
case Instruction::BitCast:
|
|
|
|
// Look past bitcasts if its operand is in the same BB.
|
|
|
|
if (InMBB)
|
2014-09-16 07:20:17 +08:00
|
|
|
return computeCallAddress(U->getOperand(0), Addr);
|
2014-07-31 12:10:40 +08:00
|
|
|
break;
|
|
|
|
case Instruction::IntToPtr:
|
|
|
|
// Look past no-op inttoptrs if its operand is in the same BB.
|
|
|
|
if (InMBB &&
|
|
|
|
TLI.getValueType(U->getOperand(0)->getType()) == TLI.getPointerTy())
|
2014-09-16 07:20:17 +08:00
|
|
|
return computeCallAddress(U->getOperand(0), Addr);
|
2014-07-31 12:10:40 +08:00
|
|
|
break;
|
|
|
|
case Instruction::PtrToInt:
|
|
|
|
// Look past no-op ptrtoints if its operand is in the same BB.
|
|
|
|
if (InMBB &&
|
|
|
|
TLI.getValueType(U->getType()) == TLI.getPointerTy())
|
2014-09-16 07:20:17 +08:00
|
|
|
return computeCallAddress(U->getOperand(0), Addr);
|
2014-07-31 12:10:40 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (const GlobalValue *GV = dyn_cast<GlobalValue>(V)) {
|
|
|
|
Addr.setGlobalValue(GV);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If all else fails, try to materialize the value in a register.
|
|
|
|
if (!Addr.getGlobalValue()) {
|
|
|
|
Addr.setReg(getRegForValue(V));
|
|
|
|
return Addr.getReg() != 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
bool AArch64FastISel::isTypeLegal(Type *Ty, MVT &VT) {
|
2014-03-29 18:18:08 +08:00
|
|
|
EVT evt = TLI.getValueType(Ty, true);
|
|
|
|
|
|
|
|
// Only handle simple types.
|
|
|
|
if (evt == MVT::Other || !evt.isSimple())
|
|
|
|
return false;
|
|
|
|
VT = evt.getSimpleVT();
|
|
|
|
|
2014-04-30 23:29:57 +08:00
|
|
|
// This is a legal type, but it's not something we handle in fast-isel.
|
|
|
|
if (VT == MVT::f128)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Handle all other legal types, i.e. a register that will directly hold this
|
2014-03-29 18:18:08 +08:00
|
|
|
// value.
|
|
|
|
return TLI.isTypeLegal(VT);
|
|
|
|
}
|
|
|
|
|
2014-09-03 06:33:53 +08:00
|
|
|
/// \brief Determine if the value type is supported by FastISel.
|
|
|
|
///
|
|
|
|
/// FastISel for AArch64 can handle more value types than are legal. This adds
|
|
|
|
/// simple value type such as i1, i8, and i16.
|
2014-09-16 05:27:54 +08:00
|
|
|
bool AArch64FastISel::isTypeSupported(Type *Ty, MVT &VT, bool IsVectorAllowed) {
|
|
|
|
if (Ty->isVectorTy() && !IsVectorAllowed)
|
2014-09-03 06:33:53 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
if (isTypeLegal(Ty, VT))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
// If this is a type than can be sign or zero-extended to a basic operation
|
|
|
|
// go ahead and accept it now.
|
|
|
|
if (VT == MVT::i1 || VT == MVT::i8 || VT == MVT::i16)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2014-08-29 08:19:21 +08:00
|
|
|
bool AArch64FastISel::isValueAvailable(const Value *V) const {
|
|
|
|
if (!isa<Instruction>(V))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
const auto *I = cast<Instruction>(V);
|
|
|
|
if (FuncInfo.MBBMap[I->getParent()] == FuncInfo.MBB)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::simplifyAddress(Address &Addr, MVT VT) {
|
2014-09-30 08:49:54 +08:00
|
|
|
unsigned ScaleFactor = getImplicitScaleFactor(VT);
|
|
|
|
if (!ScaleFactor)
|
|
|
|
return false;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
bool ImmediateOffsetNeedsLowering = false;
|
|
|
|
bool RegisterOffsetNeedsLowering = false;
|
|
|
|
int64_t Offset = Addr.getOffset();
|
|
|
|
if (((Offset < 0) || (Offset & (ScaleFactor - 1))) && !isInt<9>(Offset))
|
|
|
|
ImmediateOffsetNeedsLowering = true;
|
|
|
|
else if (Offset > 0 && !(Offset & (ScaleFactor - 1)) &&
|
|
|
|
!isUInt<12>(Offset / ScaleFactor))
|
|
|
|
ImmediateOffsetNeedsLowering = true;
|
|
|
|
|
|
|
|
// Cannot encode an offset register and an immediate offset in the same
|
|
|
|
// instruction. Fold the immediate offset into the load/store instruction and
|
|
|
|
// emit an additonal add to take care of the offset register.
|
|
|
|
if (!ImmediateOffsetNeedsLowering && Addr.getOffset() && Addr.isRegBase() &&
|
|
|
|
Addr.getOffsetReg())
|
|
|
|
RegisterOffsetNeedsLowering = true;
|
|
|
|
|
2014-08-28 05:38:33 +08:00
|
|
|
// Cannot encode zero register as base.
|
|
|
|
if (Addr.isRegBase() && Addr.getOffsetReg() && !Addr.getReg())
|
|
|
|
RegisterOffsetNeedsLowering = true;
|
|
|
|
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
// If this is a stack pointer and the offset needs to be simplified then put
|
2014-06-10 17:52:44 +08:00
|
|
|
// the alloca address into a register, set the base type back to register and
|
|
|
|
// continue. This should almost never happen.
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
if (ImmediateOffsetNeedsLowering && Addr.isFIBase()) {
|
2014-08-22 04:57:57 +08:00
|
|
|
unsigned ResultReg = createResultReg(&AArch64::GPR64spRegClass);
|
2014-06-10 17:52:44 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::ADDXri),
|
|
|
|
ResultReg)
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
.addFrameIndex(Addr.getFI())
|
|
|
|
.addImm(0)
|
|
|
|
.addImm(0);
|
2014-06-10 17:52:44 +08:00
|
|
|
Addr.setKind(Address::RegBase);
|
|
|
|
Addr.setReg(ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
if (RegisterOffsetNeedsLowering) {
|
|
|
|
unsigned ResultReg = 0;
|
2014-08-27 08:58:30 +08:00
|
|
|
if (Addr.getReg()) {
|
|
|
|
if (Addr.getExtendType() == AArch64_AM::SXTW ||
|
|
|
|
Addr.getExtendType() == AArch64_AM::UXTW )
|
|
|
|
ResultReg = emitAddSub_rx(/*UseAdd=*/true, MVT::i64, Addr.getReg(),
|
|
|
|
/*TODO:IsKill=*/false, Addr.getOffsetReg(),
|
|
|
|
/*TODO:IsKill=*/false, Addr.getExtendType(),
|
|
|
|
Addr.getShift());
|
|
|
|
else
|
|
|
|
ResultReg = emitAddSub_rs(/*UseAdd=*/true, MVT::i64, Addr.getReg(),
|
|
|
|
/*TODO:IsKill=*/false, Addr.getOffsetReg(),
|
|
|
|
/*TODO:IsKill=*/false, AArch64_AM::LSL,
|
|
|
|
Addr.getShift());
|
|
|
|
} else {
|
|
|
|
if (Addr.getExtendType() == AArch64_AM::UXTW)
|
|
|
|
ResultReg = emitLSL_ri(MVT::i64, MVT::i32, Addr.getOffsetReg(),
|
|
|
|
/*Op0IsKill=*/false, Addr.getShift(),
|
|
|
|
/*IsZExt=*/true);
|
|
|
|
else if (Addr.getExtendType() == AArch64_AM::SXTW)
|
|
|
|
ResultReg = emitLSL_ri(MVT::i64, MVT::i32, Addr.getOffsetReg(),
|
|
|
|
/*Op0IsKill=*/false, Addr.getShift(),
|
|
|
|
/*IsZExt=*/false);
|
|
|
|
else
|
|
|
|
ResultReg = emitLSL_ri(MVT::i64, MVT::i64, Addr.getOffsetReg(),
|
|
|
|
/*Op0IsKill=*/false, Addr.getShift());
|
|
|
|
}
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Addr.setReg(ResultReg);
|
|
|
|
Addr.setOffsetReg(0);
|
|
|
|
Addr.setShift(0);
|
2014-08-27 08:58:30 +08:00
|
|
|
Addr.setExtendType(AArch64_AM::InvalidShiftExtend);
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
}
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Since the offset is too large for the load/store instruction get the
|
|
|
|
// reg+offset into a register.
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
if (ImmediateOffsetNeedsLowering) {
|
2014-09-18 15:04:49 +08:00
|
|
|
unsigned ResultReg;
|
2014-09-18 13:40:47 +08:00
|
|
|
if (Addr.getReg()) {
|
|
|
|
// Try to fold the immediate into the add instruction.
|
2014-09-18 15:04:49 +08:00
|
|
|
if (Offset < 0)
|
|
|
|
ResultReg = emitAddSub_ri(/*UseAdd=*/false, MVT::i64, Addr.getReg(),
|
|
|
|
/*IsKill=*/false, -Offset);
|
|
|
|
else
|
|
|
|
ResultReg = emitAddSub_ri(/*UseAdd=*/true, MVT::i64, Addr.getReg(),
|
|
|
|
/*IsKill=*/false, Offset);
|
2014-09-18 13:40:47 +08:00
|
|
|
if (!ResultReg) {
|
|
|
|
unsigned ImmReg = fastEmit_i(MVT::i64, MVT::i64, ISD::Constant, Offset);
|
|
|
|
ResultReg = emitAddSub_rr(/*UseAdd=*/true, MVT::i64, Addr.getReg(),
|
|
|
|
/*IsKill=*/false, ImmReg, /*IsKill=*/true);
|
|
|
|
}
|
|
|
|
} else
|
2014-09-04 04:56:59 +08:00
|
|
|
ResultReg = fastEmit_i(MVT::i64, MVT::i64, ISD::Constant, Offset);
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
|
|
|
|
if (!ResultReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
Addr.setReg(ResultReg);
|
|
|
|
Addr.setOffset(0);
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
void AArch64FastISel::addLoadStoreOperands(Address &Addr,
|
2014-05-24 20:50:23 +08:00
|
|
|
const MachineInstrBuilder &MIB,
|
2014-08-09 01:24:10 +08:00
|
|
|
unsigned Flags,
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
unsigned ScaleFactor,
|
|
|
|
MachineMemOperand *MMO) {
|
|
|
|
int64_t Offset = Addr.getOffset() / ScaleFactor;
|
2014-03-29 18:18:08 +08:00
|
|
|
// Frame base works a bit differently. Handle it separately.
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
if (Addr.isFIBase()) {
|
2014-03-29 18:18:08 +08:00
|
|
|
int FI = Addr.getFI();
|
|
|
|
// FIXME: We shouldn't be using getObjectSize/getObjectAlignment. The size
|
|
|
|
// and alignment should be based on the VT.
|
2014-08-09 01:24:10 +08:00
|
|
|
MMO = FuncInfo.MF->getMachineMemOperand(
|
|
|
|
MachinePointerInfo::getFixedStack(FI, Offset), Flags,
|
|
|
|
MFI.getObjectSize(FI), MFI.getObjectAlignment(FI));
|
2014-03-29 18:18:08 +08:00
|
|
|
// Now add the rest of the operands.
|
2014-08-09 01:24:10 +08:00
|
|
|
MIB.addFrameIndex(FI).addImm(Offset);
|
2014-03-29 18:18:08 +08:00
|
|
|
} else {
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
assert(Addr.isRegBase() && "Unexpected address kind.");
|
2014-08-22 04:57:57 +08:00
|
|
|
const MCInstrDesc &II = MIB->getDesc();
|
|
|
|
unsigned Idx = (Flags & MachineMemOperand::MOStore) ? 1 : 0;
|
|
|
|
Addr.setReg(
|
|
|
|
constrainOperandRegClass(II, Addr.getReg(), II.getNumDefs()+Idx));
|
|
|
|
Addr.setOffsetReg(
|
|
|
|
constrainOperandRegClass(II, Addr.getOffsetReg(), II.getNumDefs()+Idx+1));
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
if (Addr.getOffsetReg()) {
|
|
|
|
assert(Addr.getOffset() == 0 && "Unexpected offset");
|
|
|
|
bool IsSigned = Addr.getExtendType() == AArch64_AM::SXTW ||
|
|
|
|
Addr.getExtendType() == AArch64_AM::SXTX;
|
|
|
|
MIB.addReg(Addr.getReg());
|
|
|
|
MIB.addReg(Addr.getOffsetReg());
|
|
|
|
MIB.addImm(IsSigned);
|
|
|
|
MIB.addImm(Addr.getShift() != 0);
|
|
|
|
} else {
|
|
|
|
MIB.addReg(Addr.getReg());
|
|
|
|
MIB.addImm(Offset);
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
2014-08-09 01:24:10 +08:00
|
|
|
|
|
|
|
if (MMO)
|
|
|
|
MIB.addMemOperand(MMO);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned AArch64FastISel::emitAddSub(bool UseAdd, MVT RetVT, const Value *LHS,
|
|
|
|
const Value *RHS, bool SetFlags,
|
|
|
|
bool WantResult, bool IsZExt) {
|
2014-08-20 06:29:55 +08:00
|
|
|
AArch64_AM::ShiftExtendType ExtendType = AArch64_AM::InvalidShiftExtend;
|
2014-08-21 00:34:15 +08:00
|
|
|
bool NeedExtend = false;
|
2014-08-20 06:29:55 +08:00
|
|
|
switch (RetVT.SimpleTy) {
|
2014-08-21 00:34:15 +08:00
|
|
|
default:
|
|
|
|
return 0;
|
2014-08-20 06:29:55 +08:00
|
|
|
case MVT::i1:
|
2014-08-21 00:34:15 +08:00
|
|
|
NeedExtend = true;
|
|
|
|
break;
|
2014-08-20 06:29:55 +08:00
|
|
|
case MVT::i8:
|
2014-08-21 00:34:15 +08:00
|
|
|
NeedExtend = true;
|
|
|
|
ExtendType = IsZExt ? AArch64_AM::UXTB : AArch64_AM::SXTB;
|
2014-08-20 06:29:55 +08:00
|
|
|
break;
|
|
|
|
case MVT::i16:
|
2014-08-21 00:34:15 +08:00
|
|
|
NeedExtend = true;
|
|
|
|
ExtendType = IsZExt ? AArch64_AM::UXTH : AArch64_AM::SXTH;
|
|
|
|
break;
|
|
|
|
case MVT::i32: // fall-through
|
|
|
|
case MVT::i64:
|
2014-08-20 06:29:55 +08:00
|
|
|
break;
|
|
|
|
}
|
2014-08-21 00:34:15 +08:00
|
|
|
MVT SrcVT = RetVT;
|
|
|
|
RetVT.SimpleTy = std::max(RetVT.SimpleTy, MVT::i32);
|
2014-08-20 06:29:55 +08:00
|
|
|
|
|
|
|
// Canonicalize immediates to the RHS first.
|
2014-09-03 09:38:36 +08:00
|
|
|
if (UseAdd && isa<ConstantInt>(LHS) && !isa<ConstantInt>(RHS))
|
2014-08-20 06:29:55 +08:00
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
2014-09-18 03:51:38 +08:00
|
|
|
// Canonicalize mul by power of 2 to the RHS.
|
|
|
|
if (UseAdd && LHS->hasOneUse() && isValueAvailable(LHS))
|
|
|
|
if (isMulPowOf2(LHS))
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
2014-08-20 06:29:55 +08:00
|
|
|
// Canonicalize shift immediate to the RHS.
|
2014-09-18 03:51:38 +08:00
|
|
|
if (UseAdd && LHS->hasOneUse() && isValueAvailable(LHS))
|
2014-08-20 06:29:55 +08:00
|
|
|
if (const auto *SI = dyn_cast<BinaryOperator>(LHS))
|
|
|
|
if (isa<ConstantInt>(SI->getOperand(1)))
|
|
|
|
if (SI->getOpcode() == Instruction::Shl ||
|
|
|
|
SI->getOpcode() == Instruction::LShr ||
|
|
|
|
SI->getOpcode() == Instruction::AShr )
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
|
|
|
unsigned LHSReg = getRegForValue(LHS);
|
|
|
|
if (!LHSReg)
|
|
|
|
return 0;
|
|
|
|
bool LHSIsKill = hasTrivialKill(LHS);
|
|
|
|
|
2014-08-21 00:34:15 +08:00
|
|
|
if (NeedExtend)
|
2014-09-16 07:20:17 +08:00
|
|
|
LHSReg = emitIntExt(SrcVT, LHSReg, RetVT, IsZExt);
|
2014-08-20 06:29:55 +08:00
|
|
|
|
|
|
|
unsigned ResultReg = 0;
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(RHS)) {
|
|
|
|
uint64_t Imm = IsZExt ? C->getZExtValue() : C->getSExtValue();
|
|
|
|
if (C->isNegative())
|
2014-09-03 09:38:36 +08:00
|
|
|
ResultReg = emitAddSub_ri(!UseAdd, RetVT, LHSReg, LHSIsKill, -Imm,
|
|
|
|
SetFlags, WantResult);
|
2014-08-20 06:29:55 +08:00
|
|
|
else
|
2014-09-03 09:38:36 +08:00
|
|
|
ResultReg = emitAddSub_ri(UseAdd, RetVT, LHSReg, LHSIsKill, Imm, SetFlags,
|
|
|
|
WantResult);
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
if (ResultReg)
|
|
|
|
return ResultReg;
|
|
|
|
|
2014-08-21 00:34:15 +08:00
|
|
|
// Only extend the RHS within the instruction if there is a valid extend type.
|
2014-09-18 03:51:38 +08:00
|
|
|
if (ExtendType != AArch64_AM::InvalidShiftExtend && RHS->hasOneUse() &&
|
|
|
|
isValueAvailable(RHS)) {
|
2014-08-20 06:29:55 +08:00
|
|
|
if (const auto *SI = dyn_cast<BinaryOperator>(RHS))
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(SI->getOperand(1)))
|
|
|
|
if ((SI->getOpcode() == Instruction::Shl) && (C->getZExtValue() < 4)) {
|
|
|
|
unsigned RHSReg = getRegForValue(SI->getOperand(0));
|
|
|
|
if (!RHSReg)
|
|
|
|
return 0;
|
|
|
|
bool RHSIsKill = hasTrivialKill(SI->getOperand(0));
|
2014-09-03 09:38:36 +08:00
|
|
|
return emitAddSub_rx(UseAdd, RetVT, LHSReg, LHSIsKill, RHSReg,
|
|
|
|
RHSIsKill, ExtendType, C->getZExtValue(),
|
|
|
|
SetFlags, WantResult);
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
unsigned RHSReg = getRegForValue(RHS);
|
|
|
|
if (!RHSReg)
|
|
|
|
return 0;
|
|
|
|
bool RHSIsKill = hasTrivialKill(RHS);
|
2014-09-03 09:38:36 +08:00
|
|
|
return emitAddSub_rx(UseAdd, RetVT, LHSReg, LHSIsKill, RHSReg, RHSIsKill,
|
|
|
|
ExtendType, 0, SetFlags, WantResult);
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
|
2014-09-18 03:51:38 +08:00
|
|
|
// Check if the mul can be folded into the instruction.
|
|
|
|
if (RHS->hasOneUse() && isValueAvailable(RHS))
|
|
|
|
if (isMulPowOf2(RHS)) {
|
|
|
|
const Value *MulLHS = cast<MulOperator>(RHS)->getOperand(0);
|
|
|
|
const Value *MulRHS = cast<MulOperator>(RHS)->getOperand(1);
|
|
|
|
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(MulLHS))
|
|
|
|
if (C->getValue().isPowerOf2())
|
|
|
|
std::swap(MulLHS, MulRHS);
|
|
|
|
|
|
|
|
assert(isa<ConstantInt>(MulRHS) && "Expected a ConstantInt.");
|
|
|
|
uint64_t ShiftVal = cast<ConstantInt>(MulRHS)->getValue().logBase2();
|
|
|
|
unsigned RHSReg = getRegForValue(MulLHS);
|
|
|
|
if (!RHSReg)
|
|
|
|
return 0;
|
|
|
|
bool RHSIsKill = hasTrivialKill(MulLHS);
|
|
|
|
return emitAddSub_rs(UseAdd, RetVT, LHSReg, LHSIsKill, RHSReg, RHSIsKill,
|
|
|
|
AArch64_AM::LSL, ShiftVal, SetFlags, WantResult);
|
|
|
|
}
|
|
|
|
|
2014-08-20 06:29:55 +08:00
|
|
|
// Check if the shift can be folded into the instruction.
|
2014-09-18 03:51:38 +08:00
|
|
|
if (RHS->hasOneUse() && isValueAvailable(RHS))
|
2014-08-29 08:19:21 +08:00
|
|
|
if (const auto *SI = dyn_cast<BinaryOperator>(RHS)) {
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(SI->getOperand(1))) {
|
|
|
|
AArch64_AM::ShiftExtendType ShiftType = AArch64_AM::InvalidShiftExtend;
|
|
|
|
switch (SI->getOpcode()) {
|
|
|
|
default: break;
|
|
|
|
case Instruction::Shl: ShiftType = AArch64_AM::LSL; break;
|
|
|
|
case Instruction::LShr: ShiftType = AArch64_AM::LSR; break;
|
|
|
|
case Instruction::AShr: ShiftType = AArch64_AM::ASR; break;
|
|
|
|
}
|
|
|
|
uint64_t ShiftVal = C->getZExtValue();
|
|
|
|
if (ShiftType != AArch64_AM::InvalidShiftExtend) {
|
|
|
|
unsigned RHSReg = getRegForValue(SI->getOperand(0));
|
|
|
|
if (!RHSReg)
|
|
|
|
return 0;
|
|
|
|
bool RHSIsKill = hasTrivialKill(SI->getOperand(0));
|
2014-09-03 09:38:36 +08:00
|
|
|
return emitAddSub_rs(UseAdd, RetVT, LHSReg, LHSIsKill, RHSReg,
|
|
|
|
RHSIsKill, ShiftType, ShiftVal, SetFlags,
|
|
|
|
WantResult);
|
2014-08-29 08:19:21 +08:00
|
|
|
}
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned RHSReg = getRegForValue(RHS);
|
|
|
|
if (!RHSReg)
|
|
|
|
return 0;
|
|
|
|
bool RHSIsKill = hasTrivialKill(RHS);
|
2014-08-21 00:34:15 +08:00
|
|
|
|
|
|
|
if (NeedExtend)
|
2014-09-16 07:20:17 +08:00
|
|
|
RHSReg = emitIntExt(SrcVT, RHSReg, RetVT, IsZExt);
|
2014-08-21 00:34:15 +08:00
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
return emitAddSub_rr(UseAdd, RetVT, LHSReg, LHSIsKill, RHSReg, RHSIsKill,
|
|
|
|
SetFlags, WantResult);
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned AArch64FastISel::emitAddSub_rr(bool UseAdd, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg,
|
|
|
|
bool RHSIsKill, bool SetFlags,
|
|
|
|
bool WantResult) {
|
2014-08-20 06:29:55 +08:00
|
|
|
assert(LHSReg && RHSReg && "Invalid register number.");
|
|
|
|
|
|
|
|
if (RetVT != MVT::i32 && RetVT != MVT::i64)
|
|
|
|
return 0;
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
static const unsigned OpcTable[2][2][2] = {
|
|
|
|
{ { AArch64::SUBWrr, AArch64::SUBXrr },
|
|
|
|
{ AArch64::ADDWrr, AArch64::ADDXrr } },
|
|
|
|
{ { AArch64::SUBSWrr, AArch64::SUBSXrr },
|
|
|
|
{ AArch64::ADDSWrr, AArch64::ADDSXrr } }
|
2014-08-20 06:29:55 +08:00
|
|
|
};
|
2014-09-03 09:38:36 +08:00
|
|
|
bool Is64Bit = RetVT == MVT::i64;
|
|
|
|
unsigned Opc = OpcTable[SetFlags][UseAdd][Is64Bit];
|
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
Is64Bit ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
2014-08-20 06:29:55 +08:00
|
|
|
unsigned ResultReg;
|
2014-09-03 09:38:36 +08:00
|
|
|
if (WantResult)
|
2014-08-22 04:57:57 +08:00
|
|
|
ResultReg = createResultReg(RC);
|
2014-09-03 09:38:36 +08:00
|
|
|
else
|
|
|
|
ResultReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
|
2014-08-20 06:29:55 +08:00
|
|
|
|
2014-08-22 04:57:57 +08:00
|
|
|
const MCInstrDesc &II = TII.get(Opc);
|
|
|
|
LHSReg = constrainOperandRegClass(II, LHSReg, II.getNumDefs());
|
|
|
|
RHSReg = constrainOperandRegClass(II, RHSReg, II.getNumDefs() + 1);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, II, ResultReg)
|
2014-08-20 06:29:55 +08:00
|
|
|
.addReg(LHSReg, getKillRegState(LHSIsKill))
|
|
|
|
.addReg(RHSReg, getKillRegState(RHSIsKill));
|
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned AArch64FastISel::emitAddSub_ri(bool UseAdd, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, uint64_t Imm,
|
|
|
|
bool SetFlags, bool WantResult) {
|
2014-08-20 06:29:55 +08:00
|
|
|
assert(LHSReg && "Invalid register number.");
|
|
|
|
|
|
|
|
if (RetVT != MVT::i32 && RetVT != MVT::i64)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
unsigned ShiftImm;
|
|
|
|
if (isUInt<12>(Imm))
|
|
|
|
ShiftImm = 0;
|
|
|
|
else if ((Imm & 0xfff000) == Imm) {
|
|
|
|
ShiftImm = 12;
|
|
|
|
Imm >>= 12;
|
|
|
|
} else
|
|
|
|
return 0;
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
static const unsigned OpcTable[2][2][2] = {
|
|
|
|
{ { AArch64::SUBWri, AArch64::SUBXri },
|
|
|
|
{ AArch64::ADDWri, AArch64::ADDXri } },
|
|
|
|
{ { AArch64::SUBSWri, AArch64::SUBSXri },
|
|
|
|
{ AArch64::ADDSWri, AArch64::ADDSXri } }
|
2014-08-20 06:29:55 +08:00
|
|
|
};
|
2014-09-03 09:38:36 +08:00
|
|
|
bool Is64Bit = RetVT == MVT::i64;
|
|
|
|
unsigned Opc = OpcTable[SetFlags][UseAdd][Is64Bit];
|
|
|
|
const TargetRegisterClass *RC;
|
|
|
|
if (SetFlags)
|
|
|
|
RC = Is64Bit ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
|
|
|
else
|
|
|
|
RC = Is64Bit ? &AArch64::GPR64spRegClass : &AArch64::GPR32spRegClass;
|
2014-08-20 06:29:55 +08:00
|
|
|
unsigned ResultReg;
|
2014-09-03 09:38:36 +08:00
|
|
|
if (WantResult)
|
2014-08-22 04:57:57 +08:00
|
|
|
ResultReg = createResultReg(RC);
|
2014-09-03 09:38:36 +08:00
|
|
|
else
|
|
|
|
ResultReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
|
2014-08-20 06:29:55 +08:00
|
|
|
|
2014-08-22 04:57:57 +08:00
|
|
|
const MCInstrDesc &II = TII.get(Opc);
|
|
|
|
LHSReg = constrainOperandRegClass(II, LHSReg, II.getNumDefs());
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, II, ResultReg)
|
2014-08-20 06:29:55 +08:00
|
|
|
.addReg(LHSReg, getKillRegState(LHSIsKill))
|
|
|
|
.addImm(Imm)
|
|
|
|
.addImm(getShifterImm(AArch64_AM::LSL, ShiftImm));
|
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned AArch64FastISel::emitAddSub_rs(bool UseAdd, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg,
|
|
|
|
bool RHSIsKill,
|
2014-08-27 08:58:30 +08:00
|
|
|
AArch64_AM::ShiftExtendType ShiftType,
|
2014-09-03 09:38:36 +08:00
|
|
|
uint64_t ShiftImm, bool SetFlags,
|
|
|
|
bool WantResult) {
|
2014-08-20 06:29:55 +08:00
|
|
|
assert(LHSReg && RHSReg && "Invalid register number.");
|
|
|
|
|
|
|
|
if (RetVT != MVT::i32 && RetVT != MVT::i64)
|
|
|
|
return 0;
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
static const unsigned OpcTable[2][2][2] = {
|
|
|
|
{ { AArch64::SUBWrs, AArch64::SUBXrs },
|
|
|
|
{ AArch64::ADDWrs, AArch64::ADDXrs } },
|
|
|
|
{ { AArch64::SUBSWrs, AArch64::SUBSXrs },
|
|
|
|
{ AArch64::ADDSWrs, AArch64::ADDSXrs } }
|
2014-08-20 06:29:55 +08:00
|
|
|
};
|
2014-09-03 09:38:36 +08:00
|
|
|
bool Is64Bit = RetVT == MVT::i64;
|
|
|
|
unsigned Opc = OpcTable[SetFlags][UseAdd][Is64Bit];
|
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
Is64Bit ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
2014-08-20 06:29:55 +08:00
|
|
|
unsigned ResultReg;
|
2014-09-03 09:38:36 +08:00
|
|
|
if (WantResult)
|
2014-08-22 04:57:57 +08:00
|
|
|
ResultReg = createResultReg(RC);
|
2014-09-03 09:38:36 +08:00
|
|
|
else
|
|
|
|
ResultReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
|
2014-08-20 06:29:55 +08:00
|
|
|
|
2014-08-22 04:57:57 +08:00
|
|
|
const MCInstrDesc &II = TII.get(Opc);
|
|
|
|
LHSReg = constrainOperandRegClass(II, LHSReg, II.getNumDefs());
|
|
|
|
RHSReg = constrainOperandRegClass(II, RHSReg, II.getNumDefs() + 1);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, II, ResultReg)
|
2014-08-20 06:29:55 +08:00
|
|
|
.addReg(LHSReg, getKillRegState(LHSIsKill))
|
|
|
|
.addReg(RHSReg, getKillRegState(RHSIsKill))
|
|
|
|
.addImm(getShifterImm(ShiftType, ShiftImm));
|
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned AArch64FastISel::emitAddSub_rx(bool UseAdd, MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg,
|
|
|
|
bool RHSIsKill,
|
2014-08-27 08:58:30 +08:00
|
|
|
AArch64_AM::ShiftExtendType ExtType,
|
2014-09-03 09:38:36 +08:00
|
|
|
uint64_t ShiftImm, bool SetFlags,
|
|
|
|
bool WantResult) {
|
2014-08-27 08:58:30 +08:00
|
|
|
assert(LHSReg && RHSReg && "Invalid register number.");
|
|
|
|
|
|
|
|
if (RetVT != MVT::i32 && RetVT != MVT::i64)
|
|
|
|
return 0;
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
static const unsigned OpcTable[2][2][2] = {
|
|
|
|
{ { AArch64::SUBWrx, AArch64::SUBXrx },
|
|
|
|
{ AArch64::ADDWrx, AArch64::ADDXrx } },
|
|
|
|
{ { AArch64::SUBSWrx, AArch64::SUBSXrx },
|
|
|
|
{ AArch64::ADDSWrx, AArch64::ADDSXrx } }
|
2014-08-27 08:58:30 +08:00
|
|
|
};
|
2014-09-03 09:38:36 +08:00
|
|
|
bool Is64Bit = RetVT == MVT::i64;
|
|
|
|
unsigned Opc = OpcTable[SetFlags][UseAdd][Is64Bit];
|
|
|
|
const TargetRegisterClass *RC = nullptr;
|
|
|
|
if (SetFlags)
|
|
|
|
RC = Is64Bit ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
|
|
|
else
|
|
|
|
RC = Is64Bit ? &AArch64::GPR64spRegClass : &AArch64::GPR32spRegClass;
|
2014-08-20 06:29:55 +08:00
|
|
|
unsigned ResultReg;
|
2014-09-03 09:38:36 +08:00
|
|
|
if (WantResult)
|
2014-08-22 04:57:57 +08:00
|
|
|
ResultReg = createResultReg(RC);
|
2014-09-03 09:38:36 +08:00
|
|
|
else
|
|
|
|
ResultReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
|
2014-08-20 06:29:55 +08:00
|
|
|
|
2014-08-22 04:57:57 +08:00
|
|
|
const MCInstrDesc &II = TII.get(Opc);
|
|
|
|
LHSReg = constrainOperandRegClass(II, LHSReg, II.getNumDefs());
|
|
|
|
RHSReg = constrainOperandRegClass(II, RHSReg, II.getNumDefs() + 1);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, II, ResultReg)
|
2014-08-20 06:29:55 +08:00
|
|
|
.addReg(LHSReg, getKillRegState(LHSIsKill))
|
|
|
|
.addReg(RHSReg, getKillRegState(RHSIsKill))
|
|
|
|
.addImm(getArithExtendImm(ExtType, ShiftImm));
|
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AArch64FastISel::emitCmp(const Value *LHS, const Value *RHS, bool IsZExt) {
|
|
|
|
Type *Ty = LHS->getType();
|
|
|
|
EVT EVT = TLI.getValueType(Ty, true);
|
|
|
|
if (!EVT.isSimple())
|
|
|
|
return false;
|
|
|
|
MVT VT = EVT.getSimpleVT();
|
|
|
|
|
|
|
|
switch (VT.SimpleTy) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case MVT::i1:
|
|
|
|
case MVT::i8:
|
|
|
|
case MVT::i16:
|
|
|
|
case MVT::i32:
|
|
|
|
case MVT::i64:
|
|
|
|
return emitICmp(VT, LHS, RHS, IsZExt);
|
|
|
|
case MVT::f32:
|
|
|
|
case MVT::f64:
|
|
|
|
return emitFCmp(VT, LHS, RHS);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool AArch64FastISel::emitICmp(MVT RetVT, const Value *LHS, const Value *RHS,
|
|
|
|
bool IsZExt) {
|
2014-09-03 09:38:36 +08:00
|
|
|
return emitSub(RetVT, LHS, RHS, /*SetFlags=*/true, /*WantResult=*/false,
|
|
|
|
IsZExt) != 0;
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bool AArch64FastISel::emitICmp_ri(MVT RetVT, unsigned LHSReg, bool LHSIsKill,
|
|
|
|
uint64_t Imm) {
|
2014-09-03 09:38:36 +08:00
|
|
|
return emitAddSub_ri(/*UseAdd=*/false, RetVT, LHSReg, LHSIsKill, Imm,
|
|
|
|
/*SetFlags=*/true, /*WantResult=*/false) != 0;
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bool AArch64FastISel::emitFCmp(MVT RetVT, const Value *LHS, const Value *RHS) {
|
|
|
|
if (RetVT != MVT::f32 && RetVT != MVT::f64)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Check to see if the 2nd operand is a constant that we can encode directly
|
|
|
|
// in the compare.
|
|
|
|
bool UseImm = false;
|
|
|
|
if (const auto *CFP = dyn_cast<ConstantFP>(RHS))
|
|
|
|
if (CFP->isZero() && !CFP->isNegative())
|
|
|
|
UseImm = true;
|
|
|
|
|
|
|
|
unsigned LHSReg = getRegForValue(LHS);
|
|
|
|
if (!LHSReg)
|
|
|
|
return false;
|
|
|
|
bool LHSIsKill = hasTrivialKill(LHS);
|
|
|
|
|
|
|
|
if (UseImm) {
|
|
|
|
unsigned Opc = (RetVT == MVT::f64) ? AArch64::FCMPDri : AArch64::FCMPSri;
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(Opc))
|
|
|
|
.addReg(LHSReg, getKillRegState(LHSIsKill));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned RHSReg = getRegForValue(RHS);
|
|
|
|
if (!RHSReg)
|
|
|
|
return false;
|
|
|
|
bool RHSIsKill = hasTrivialKill(RHS);
|
|
|
|
|
|
|
|
unsigned Opc = (RetVT == MVT::f64) ? AArch64::FCMPDrr : AArch64::FCMPSrr;
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(Opc))
|
|
|
|
.addReg(LHSReg, getKillRegState(LHSIsKill))
|
|
|
|
.addReg(RHSReg, getKillRegState(RHSIsKill));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned AArch64FastISel::emitAdd(MVT RetVT, const Value *LHS, const Value *RHS,
|
|
|
|
bool SetFlags, bool WantResult, bool IsZExt) {
|
|
|
|
return emitAddSub(/*UseAdd=*/true, RetVT, LHS, RHS, SetFlags, WantResult,
|
|
|
|
IsZExt);
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned AArch64FastISel::emitSub(MVT RetVT, const Value *LHS, const Value *RHS,
|
|
|
|
bool SetFlags, bool WantResult, bool IsZExt) {
|
|
|
|
return emitAddSub(/*UseAdd=*/false, RetVT, LHS, RHS, SetFlags, WantResult,
|
|
|
|
IsZExt);
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned AArch64FastISel::emitSubs_rr(MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg,
|
|
|
|
bool RHSIsKill, bool WantResult) {
|
2014-09-03 09:38:36 +08:00
|
|
|
return emitAddSub_rr(/*UseAdd=*/false, RetVT, LHSReg, LHSIsKill, RHSReg,
|
|
|
|
RHSIsKill, /*SetFlags=*/true, WantResult);
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned AArch64FastISel::emitSubs_rs(MVT RetVT, unsigned LHSReg,
|
|
|
|
bool LHSIsKill, unsigned RHSReg,
|
|
|
|
bool RHSIsKill,
|
|
|
|
AArch64_AM::ShiftExtendType ShiftType,
|
|
|
|
uint64_t ShiftImm, bool WantResult) {
|
2014-09-03 09:38:36 +08:00
|
|
|
return emitAddSub_rs(/*UseAdd=*/false, RetVT, LHSReg, LHSIsKill, RHSReg,
|
|
|
|
RHSIsKill, ShiftType, ShiftImm, /*SetFlags=*/true,
|
|
|
|
WantResult);
|
2014-08-20 06:29:55 +08:00
|
|
|
}
|
|
|
|
|
2014-09-04 09:29:18 +08:00
|
|
|
unsigned AArch64FastISel::emitLogicalOp(unsigned ISDOpc, MVT RetVT,
|
|
|
|
const Value *LHS, const Value *RHS) {
|
|
|
|
// Canonicalize immediates to the RHS first.
|
|
|
|
if (isa<ConstantInt>(LHS) && !isa<ConstantInt>(RHS))
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
2014-09-18 03:51:38 +08:00
|
|
|
// Canonicalize mul by power-of-2 to the RHS.
|
|
|
|
if (LHS->hasOneUse() && isValueAvailable(LHS))
|
|
|
|
if (isMulPowOf2(LHS))
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
2014-09-04 09:29:18 +08:00
|
|
|
// Canonicalize shift immediate to the RHS.
|
2014-09-18 03:51:38 +08:00
|
|
|
if (LHS->hasOneUse() && isValueAvailable(LHS))
|
|
|
|
if (const auto *SI = dyn_cast<ShlOperator>(LHS))
|
2014-09-04 09:29:18 +08:00
|
|
|
if (isa<ConstantInt>(SI->getOperand(1)))
|
2014-09-18 03:51:38 +08:00
|
|
|
std::swap(LHS, RHS);
|
2014-09-04 09:29:18 +08:00
|
|
|
|
|
|
|
unsigned LHSReg = getRegForValue(LHS);
|
|
|
|
if (!LHSReg)
|
|
|
|
return 0;
|
|
|
|
bool LHSIsKill = hasTrivialKill(LHS);
|
|
|
|
|
|
|
|
unsigned ResultReg = 0;
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(RHS)) {
|
|
|
|
uint64_t Imm = C->getZExtValue();
|
|
|
|
ResultReg = emitLogicalOp_ri(ISDOpc, RetVT, LHSReg, LHSIsKill, Imm);
|
|
|
|
}
|
|
|
|
if (ResultReg)
|
|
|
|
return ResultReg;
|
|
|
|
|
2014-09-18 03:51:38 +08:00
|
|
|
// Check if the mul can be folded into the instruction.
|
|
|
|
if (RHS->hasOneUse() && isValueAvailable(RHS))
|
|
|
|
if (isMulPowOf2(RHS)) {
|
|
|
|
const Value *MulLHS = cast<MulOperator>(RHS)->getOperand(0);
|
|
|
|
const Value *MulRHS = cast<MulOperator>(RHS)->getOperand(1);
|
|
|
|
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(MulLHS))
|
|
|
|
if (C->getValue().isPowerOf2())
|
|
|
|
std::swap(MulLHS, MulRHS);
|
|
|
|
|
|
|
|
assert(isa<ConstantInt>(MulRHS) && "Expected a ConstantInt.");
|
|
|
|
uint64_t ShiftVal = cast<ConstantInt>(MulRHS)->getValue().logBase2();
|
|
|
|
|
|
|
|
unsigned RHSReg = getRegForValue(MulLHS);
|
|
|
|
if (!RHSReg)
|
|
|
|
return 0;
|
|
|
|
bool RHSIsKill = hasTrivialKill(MulLHS);
|
|
|
|
return emitLogicalOp_rs(ISDOpc, RetVT, LHSReg, LHSIsKill, RHSReg,
|
|
|
|
RHSIsKill, ShiftVal);
|
|
|
|
}
|
|
|
|
|
2014-09-04 09:29:18 +08:00
|
|
|
// Check if the shift can be folded into the instruction.
|
2014-09-18 03:51:38 +08:00
|
|
|
if (RHS->hasOneUse() && isValueAvailable(RHS))
|
|
|
|
if (const auto *SI = dyn_cast<ShlOperator>(RHS))
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(SI->getOperand(1))) {
|
|
|
|
uint64_t ShiftVal = C->getZExtValue();
|
|
|
|
unsigned RHSReg = getRegForValue(SI->getOperand(0));
|
|
|
|
if (!RHSReg)
|
|
|
|
return 0;
|
|
|
|
bool RHSIsKill = hasTrivialKill(SI->getOperand(0));
|
|
|
|
return emitLogicalOp_rs(ISDOpc, RetVT, LHSReg, LHSIsKill, RHSReg,
|
|
|
|
RHSIsKill, ShiftVal);
|
|
|
|
}
|
2014-09-04 09:29:18 +08:00
|
|
|
|
|
|
|
unsigned RHSReg = getRegForValue(RHS);
|
|
|
|
if (!RHSReg)
|
|
|
|
return 0;
|
|
|
|
bool RHSIsKill = hasTrivialKill(RHS);
|
|
|
|
|
2014-09-14 07:46:28 +08:00
|
|
|
MVT VT = std::max(MVT::i32, RetVT.SimpleTy);
|
|
|
|
ResultReg = fastEmit_rr(VT, VT, ISDOpc, LHSReg, LHSIsKill, RHSReg, RHSIsKill);
|
|
|
|
if (RetVT >= MVT::i8 && RetVT <= MVT::i16) {
|
|
|
|
uint64_t Mask = (RetVT == MVT::i8) ? 0xff : 0xffff;
|
|
|
|
ResultReg = emitAnd_ri(MVT::i32, ResultReg, /*IsKill=*/true, Mask);
|
|
|
|
}
|
|
|
|
return ResultReg;
|
2014-09-04 09:29:18 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned AArch64FastISel::emitLogicalOp_ri(unsigned ISDOpc, MVT RetVT,
|
|
|
|
unsigned LHSReg, bool LHSIsKill,
|
|
|
|
uint64_t Imm) {
|
|
|
|
assert((ISD::AND + 1 == ISD::OR) && (ISD::AND + 2 == ISD::XOR) &&
|
|
|
|
"ISD nodes are not consecutive!");
|
|
|
|
static const unsigned OpcTable[3][2] = {
|
|
|
|
{ AArch64::ANDWri, AArch64::ANDXri },
|
|
|
|
{ AArch64::ORRWri, AArch64::ORRXri },
|
|
|
|
{ AArch64::EORWri, AArch64::EORXri }
|
|
|
|
};
|
|
|
|
const TargetRegisterClass *RC;
|
|
|
|
unsigned Opc;
|
|
|
|
unsigned RegSize;
|
2014-08-22 02:02:25 +08:00
|
|
|
switch (RetVT.SimpleTy) {
|
|
|
|
default:
|
|
|
|
return 0;
|
2014-09-14 07:46:28 +08:00
|
|
|
case MVT::i1:
|
|
|
|
case MVT::i8:
|
|
|
|
case MVT::i16:
|
2014-09-04 09:29:18 +08:00
|
|
|
case MVT::i32: {
|
|
|
|
unsigned Idx = ISDOpc - ISD::AND;
|
|
|
|
Opc = OpcTable[Idx][0];
|
2014-08-22 02:02:25 +08:00
|
|
|
RC = &AArch64::GPR32spRegClass;
|
|
|
|
RegSize = 32;
|
|
|
|
break;
|
2014-09-04 09:29:18 +08:00
|
|
|
}
|
2014-08-22 02:02:25 +08:00
|
|
|
case MVT::i64:
|
2014-09-04 09:29:18 +08:00
|
|
|
Opc = OpcTable[ISDOpc - ISD::AND][1];
|
2014-08-22 02:02:25 +08:00
|
|
|
RC = &AArch64::GPR64spRegClass;
|
|
|
|
RegSize = 64;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!AArch64_AM::isLogicalImmediate(Imm, RegSize))
|
|
|
|
return 0;
|
|
|
|
|
2014-09-14 07:46:28 +08:00
|
|
|
unsigned ResultReg =
|
|
|
|
fastEmitInst_ri(Opc, RC, LHSReg, LHSIsKill,
|
|
|
|
AArch64_AM::encodeLogicalImmediate(Imm, RegSize));
|
|
|
|
if (RetVT >= MVT::i8 && RetVT <= MVT::i16 && ISDOpc != ISD::AND) {
|
|
|
|
uint64_t Mask = (RetVT == MVT::i8) ? 0xff : 0xffff;
|
|
|
|
ResultReg = emitAnd_ri(MVT::i32, ResultReg, /*IsKill=*/true, Mask);
|
|
|
|
}
|
|
|
|
return ResultReg;
|
2014-08-22 02:02:25 +08:00
|
|
|
}
|
|
|
|
|
2014-09-04 09:29:18 +08:00
|
|
|
unsigned AArch64FastISel::emitLogicalOp_rs(unsigned ISDOpc, MVT RetVT,
|
|
|
|
unsigned LHSReg, bool LHSIsKill,
|
|
|
|
unsigned RHSReg, bool RHSIsKill,
|
|
|
|
uint64_t ShiftImm) {
|
|
|
|
assert((ISD::AND + 1 == ISD::OR) && (ISD::AND + 2 == ISD::XOR) &&
|
|
|
|
"ISD nodes are not consecutive!");
|
|
|
|
static const unsigned OpcTable[3][2] = {
|
|
|
|
{ AArch64::ANDWrs, AArch64::ANDXrs },
|
|
|
|
{ AArch64::ORRWrs, AArch64::ORRXrs },
|
|
|
|
{ AArch64::EORWrs, AArch64::EORXrs }
|
|
|
|
};
|
|
|
|
const TargetRegisterClass *RC;
|
|
|
|
unsigned Opc;
|
|
|
|
switch (RetVT.SimpleTy) {
|
2014-09-14 07:46:28 +08:00
|
|
|
default:
|
|
|
|
return 0;
|
|
|
|
case MVT::i1:
|
|
|
|
case MVT::i8:
|
|
|
|
case MVT::i16:
|
|
|
|
case MVT::i32:
|
|
|
|
Opc = OpcTable[ISDOpc - ISD::AND][0];
|
|
|
|
RC = &AArch64::GPR32RegClass;
|
|
|
|
break;
|
|
|
|
case MVT::i64:
|
|
|
|
Opc = OpcTable[ISDOpc - ISD::AND][1];
|
|
|
|
RC = &AArch64::GPR64RegClass;
|
|
|
|
break;
|
2014-09-04 09:29:18 +08:00
|
|
|
}
|
2014-09-14 07:46:28 +08:00
|
|
|
unsigned ResultReg =
|
|
|
|
fastEmitInst_rri(Opc, RC, LHSReg, LHSIsKill, RHSReg, RHSIsKill,
|
|
|
|
AArch64_AM::getShifterImm(AArch64_AM::LSL, ShiftImm));
|
|
|
|
if (RetVT >= MVT::i8 && RetVT <= MVT::i16) {
|
|
|
|
uint64_t Mask = (RetVT == MVT::i8) ? 0xff : 0xffff;
|
|
|
|
ResultReg = emitAnd_ri(MVT::i32, ResultReg, /*IsKill=*/true, Mask);
|
|
|
|
}
|
|
|
|
return ResultReg;
|
2014-09-04 09:29:18 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned AArch64FastISel::emitAnd_ri(MVT RetVT, unsigned LHSReg, bool LHSIsKill,
|
|
|
|
uint64_t Imm) {
|
|
|
|
return emitLogicalOp_ri(ISD::AND, RetVT, LHSReg, LHSIsKill, Imm);
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::emitLoad(MVT VT, unsigned &ResultReg, Address Addr,
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
MachineMemOperand *MMO) {
|
|
|
|
// Simplify this down to something we can handle.
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!simplifyAddress(Addr, VT))
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-30 08:49:54 +08:00
|
|
|
unsigned ScaleFactor = getImplicitScaleFactor(VT);
|
|
|
|
if (!ScaleFactor)
|
|
|
|
llvm_unreachable("Unexpected value type.");
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Negative offsets require unscaled, 9-bit, signed immediate offsets.
|
|
|
|
// Otherwise, we try using scaled, 12-bit, unsigned immediate offsets.
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
bool UseScaled = true;
|
|
|
|
if ((Addr.getOffset() < 0) || (Addr.getOffset() & (ScaleFactor - 1))) {
|
|
|
|
UseScaled = false;
|
|
|
|
ScaleFactor = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const unsigned OpcTable[4][6] = {
|
|
|
|
{ AArch64::LDURBBi, AArch64::LDURHHi, AArch64::LDURWi, AArch64::LDURXi,
|
|
|
|
AArch64::LDURSi, AArch64::LDURDi },
|
|
|
|
{ AArch64::LDRBBui, AArch64::LDRHHui, AArch64::LDRWui, AArch64::LDRXui,
|
|
|
|
AArch64::LDRSui, AArch64::LDRDui },
|
|
|
|
{ AArch64::LDRBBroX, AArch64::LDRHHroX, AArch64::LDRWroX, AArch64::LDRXroX,
|
|
|
|
AArch64::LDRSroX, AArch64::LDRDroX },
|
|
|
|
{ AArch64::LDRBBroW, AArch64::LDRHHroW, AArch64::LDRWroW, AArch64::LDRXroW,
|
|
|
|
AArch64::LDRSroW, AArch64::LDRDroW }
|
|
|
|
};
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
unsigned Opc;
|
|
|
|
const TargetRegisterClass *RC;
|
|
|
|
bool VTIsi1 = false;
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
bool UseRegOffset = Addr.isRegBase() && !Addr.getOffset() && Addr.getReg() &&
|
|
|
|
Addr.getOffsetReg();
|
|
|
|
unsigned Idx = UseRegOffset ? 2 : UseScaled ? 1 : 0;
|
|
|
|
if (Addr.getExtendType() == AArch64_AM::UXTW ||
|
|
|
|
Addr.getExtendType() == AArch64_AM::SXTW)
|
|
|
|
Idx++;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
switch (VT.SimpleTy) {
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
default: llvm_unreachable("Unexpected value type.");
|
|
|
|
case MVT::i1: VTIsi1 = true; // Intentional fall-through.
|
|
|
|
case MVT::i8: Opc = OpcTable[Idx][0]; RC = &AArch64::GPR32RegClass; break;
|
|
|
|
case MVT::i16: Opc = OpcTable[Idx][1]; RC = &AArch64::GPR32RegClass; break;
|
|
|
|
case MVT::i32: Opc = OpcTable[Idx][2]; RC = &AArch64::GPR32RegClass; break;
|
|
|
|
case MVT::i64: Opc = OpcTable[Idx][3]; RC = &AArch64::GPR64RegClass; break;
|
|
|
|
case MVT::f32: Opc = OpcTable[Idx][4]; RC = &AArch64::FPR32RegClass; break;
|
|
|
|
case MVT::f64: Opc = OpcTable[Idx][5]; RC = &AArch64::FPR64RegClass; break;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Create the base instruction, then add the operands.
|
|
|
|
ResultReg = createResultReg(RC);
|
|
|
|
MachineInstrBuilder MIB = BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(Opc), ResultReg);
|
2014-09-16 07:20:17 +08:00
|
|
|
addLoadStoreOperands(Addr, MIB, MachineMemOperand::MOLoad, ScaleFactor, MMO);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Loading an i1 requires special handling.
|
|
|
|
if (VTIsi1) {
|
2014-09-04 09:29:18 +08:00
|
|
|
unsigned ANDReg = emitAnd_ri(MVT::i32, ResultReg, /*IsKill=*/true, 1);
|
2014-08-22 02:02:25 +08:00
|
|
|
assert(ANDReg && "Unexpected AND instruction emission failure.");
|
2014-03-29 18:18:08 +08:00
|
|
|
ResultReg = ANDReg;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
bool AArch64FastISel::selectAddSub(const Instruction *I) {
|
|
|
|
MVT VT;
|
2014-09-16 05:27:56 +08:00
|
|
|
if (!isTypeSupported(I->getType(), VT, /*IsVectorAllowed=*/true))
|
2014-09-03 09:38:36 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-16 05:27:56 +08:00
|
|
|
if (VT.isVector())
|
|
|
|
return selectOperator(I, I->getOpcode());
|
|
|
|
|
2014-09-03 09:38:36 +08:00
|
|
|
unsigned ResultReg;
|
2014-09-16 05:27:56 +08:00
|
|
|
switch (I->getOpcode()) {
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unexpected instruction.");
|
|
|
|
case Instruction::Add:
|
2014-09-03 09:38:36 +08:00
|
|
|
ResultReg = emitAdd(VT, I->getOperand(0), I->getOperand(1));
|
2014-09-16 05:27:56 +08:00
|
|
|
break;
|
|
|
|
case Instruction::Sub:
|
2014-09-03 09:38:36 +08:00
|
|
|
ResultReg = emitSub(VT, I->getOperand(0), I->getOperand(1));
|
2014-09-16 05:27:56 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
2014-09-03 09:38:36 +08:00
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-09-03 09:38:36 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 05:27:56 +08:00
|
|
|
bool AArch64FastISel::selectLogicalOp(const Instruction *I) {
|
2014-09-04 09:29:18 +08:00
|
|
|
MVT VT;
|
2014-09-16 05:27:56 +08:00
|
|
|
if (!isTypeSupported(I->getType(), VT, /*IsVectorAllowed=*/true))
|
2014-09-04 09:29:18 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-16 05:27:56 +08:00
|
|
|
if (VT.isVector())
|
|
|
|
return selectOperator(I, I->getOpcode());
|
|
|
|
|
|
|
|
unsigned ResultReg;
|
|
|
|
switch (I->getOpcode()) {
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unexpected instruction.");
|
|
|
|
case Instruction::And:
|
|
|
|
ResultReg = emitLogicalOp(ISD::AND, VT, I->getOperand(0), I->getOperand(1));
|
|
|
|
break;
|
|
|
|
case Instruction::Or:
|
|
|
|
ResultReg = emitLogicalOp(ISD::OR, VT, I->getOperand(0), I->getOperand(1));
|
|
|
|
break;
|
|
|
|
case Instruction::Xor:
|
|
|
|
ResultReg = emitLogicalOp(ISD::XOR, VT, I->getOperand(0), I->getOperand(1));
|
|
|
|
break;
|
|
|
|
}
|
2014-09-04 09:29:18 +08:00
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
updateValueMap(I, ResultReg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectLoad(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
MVT VT;
|
|
|
|
// Verify we have a legal type before going any further. Currently, we handle
|
|
|
|
// simple types that will directly fit in a register (i32/f32/i64/f64) or
|
|
|
|
// those that can be sign or zero-extended to a basic operation (i1/i8/i16).
|
2014-09-16 05:27:54 +08:00
|
|
|
if (!isTypeSupported(I->getType(), VT, /*IsVectorAllowed=*/true) ||
|
|
|
|
cast<LoadInst>(I)->isAtomic())
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// See if we can handle this address.
|
|
|
|
Address Addr;
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!computeAddress(I->getOperand(0), Addr, I->getType()))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned ResultReg;
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!emitLoad(VT, ResultReg, Addr, createMachineMemOperandFor(I)))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::emitStore(MVT VT, unsigned SrcReg, Address Addr,
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
MachineMemOperand *MMO) {
|
|
|
|
// Simplify this down to something we can handle.
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!simplifyAddress(Addr, VT))
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-30 08:49:54 +08:00
|
|
|
unsigned ScaleFactor = getImplicitScaleFactor(VT);
|
|
|
|
if (!ScaleFactor)
|
|
|
|
llvm_unreachable("Unexpected value type.");
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Negative offsets require unscaled, 9-bit, signed immediate offsets.
|
|
|
|
// Otherwise, we try using scaled, 12-bit, unsigned immediate offsets.
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
bool UseScaled = true;
|
|
|
|
if ((Addr.getOffset() < 0) || (Addr.getOffset() & (ScaleFactor - 1))) {
|
|
|
|
UseScaled = false;
|
|
|
|
ScaleFactor = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const unsigned OpcTable[4][6] = {
|
|
|
|
{ AArch64::STURBBi, AArch64::STURHHi, AArch64::STURWi, AArch64::STURXi,
|
|
|
|
AArch64::STURSi, AArch64::STURDi },
|
|
|
|
{ AArch64::STRBBui, AArch64::STRHHui, AArch64::STRWui, AArch64::STRXui,
|
|
|
|
AArch64::STRSui, AArch64::STRDui },
|
|
|
|
{ AArch64::STRBBroX, AArch64::STRHHroX, AArch64::STRWroX, AArch64::STRXroX,
|
|
|
|
AArch64::STRSroX, AArch64::STRDroX },
|
|
|
|
{ AArch64::STRBBroW, AArch64::STRHHroW, AArch64::STRWroW, AArch64::STRXroW,
|
|
|
|
AArch64::STRSroW, AArch64::STRDroW }
|
|
|
|
};
|
|
|
|
|
|
|
|
unsigned Opc;
|
2014-03-29 18:18:08 +08:00
|
|
|
bool VTIsi1 = false;
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
bool UseRegOffset = Addr.isRegBase() && !Addr.getOffset() && Addr.getReg() &&
|
|
|
|
Addr.getOffsetReg();
|
|
|
|
unsigned Idx = UseRegOffset ? 2 : UseScaled ? 1 : 0;
|
|
|
|
if (Addr.getExtendType() == AArch64_AM::UXTW ||
|
|
|
|
Addr.getExtendType() == AArch64_AM::SXTW)
|
|
|
|
Idx++;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
switch (VT.SimpleTy) {
|
Reapply [FastISel][AArch64] Add support for more addressing modes (r215597).
Note: This was originally reverted to track down a buildbot error. Reapply
without any modifications.
Original commit message:
FastISel didn't take much advantage of the different addressing modes available
to it on AArch64. This commit allows the ComputeAddress method to recognize more
addressing modes that allows shifts and sign-/zero-extensions to be folded into
the memory operation itself.
For Example:
lsl x1, x1, #3 --> ldr x0, [x0, x1, lsl #3]
ldr x0, [x0, x1]
sxtw x1, w1
lsl x1, x1, #3 --> ldr x0, [x0, x1, sxtw #3]
ldr x0, [x0, x1]
llvm-svn: 216013
2014-08-20 03:44:17 +08:00
|
|
|
default: llvm_unreachable("Unexpected value type.");
|
|
|
|
case MVT::i1: VTIsi1 = true;
|
|
|
|
case MVT::i8: Opc = OpcTable[Idx][0]; break;
|
|
|
|
case MVT::i16: Opc = OpcTable[Idx][1]; break;
|
|
|
|
case MVT::i32: Opc = OpcTable[Idx][2]; break;
|
|
|
|
case MVT::i64: Opc = OpcTable[Idx][3]; break;
|
|
|
|
case MVT::f32: Opc = OpcTable[Idx][4]; break;
|
|
|
|
case MVT::f64: Opc = OpcTable[Idx][5]; break;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Storing an i1 requires special handling.
|
2014-08-28 05:04:52 +08:00
|
|
|
if (VTIsi1 && SrcReg != AArch64::WZR) {
|
2014-09-04 09:29:18 +08:00
|
|
|
unsigned ANDReg = emitAnd_ri(MVT::i32, SrcReg, /*TODO:IsKill=*/false, 1);
|
2014-08-22 02:02:25 +08:00
|
|
|
assert(ANDReg && "Unexpected AND instruction emission failure.");
|
2014-03-29 18:18:08 +08:00
|
|
|
SrcReg = ANDReg;
|
|
|
|
}
|
|
|
|
// Create the base instruction, then add the operands.
|
2014-08-22 04:57:57 +08:00
|
|
|
const MCInstrDesc &II = TII.get(Opc);
|
|
|
|
SrcReg = constrainOperandRegClass(II, SrcReg, II.getNumDefs());
|
|
|
|
MachineInstrBuilder MIB =
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, II).addReg(SrcReg);
|
2014-09-16 07:20:17 +08:00
|
|
|
addLoadStoreOperands(Addr, MIB, MachineMemOperand::MOStore, ScaleFactor, MMO);
|
2014-08-09 01:24:10 +08:00
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectStore(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
MVT VT;
|
2014-08-28 05:04:52 +08:00
|
|
|
const Value *Op0 = I->getOperand(0);
|
2014-03-29 18:18:08 +08:00
|
|
|
// Verify we have a legal type before going any further. Currently, we handle
|
|
|
|
// simple types that will directly fit in a register (i32/f32/i64/f64) or
|
|
|
|
// those that can be sign or zero-extended to a basic operation (i1/i8/i16).
|
2014-09-16 05:27:54 +08:00
|
|
|
if (!isTypeSupported(Op0->getType(), VT, /*IsVectorAllowed=*/true) ||
|
2014-03-29 18:18:08 +08:00
|
|
|
cast<StoreInst>(I)->isAtomic())
|
|
|
|
return false;
|
|
|
|
|
2014-08-28 05:04:52 +08:00
|
|
|
// Get the value to be stored into a register. Use the zero register directly
|
2014-08-28 05:40:50 +08:00
|
|
|
// when possible to avoid an unnecessary copy and a wasted register.
|
2014-08-28 05:04:52 +08:00
|
|
|
unsigned SrcReg = 0;
|
|
|
|
if (const auto *CI = dyn_cast<ConstantInt>(Op0)) {
|
|
|
|
if (CI->isZero())
|
|
|
|
SrcReg = (VT == MVT::i64) ? AArch64::XZR : AArch64::WZR;
|
|
|
|
} else if (const auto *CF = dyn_cast<ConstantFP>(Op0)) {
|
|
|
|
if (CF->isZero() && !CF->isNegative()) {
|
|
|
|
VT = MVT::getIntegerVT(VT.getSizeInBits());
|
|
|
|
SrcReg = (VT == MVT::i64) ? AArch64::XZR : AArch64::WZR;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!SrcReg)
|
|
|
|
SrcReg = getRegForValue(Op0);
|
|
|
|
|
|
|
|
if (!SrcReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// See if we can handle this address.
|
|
|
|
Address Addr;
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!computeAddress(I->getOperand(1), Addr, I->getOperand(0)->getType()))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!emitStore(VT, SrcReg, Addr, createMachineMemOperandFor(I)))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
static AArch64CC::CondCode getCompareCC(CmpInst::Predicate Pred) {
|
2014-03-29 18:18:08 +08:00
|
|
|
switch (Pred) {
|
|
|
|
case CmpInst::FCMP_ONE:
|
|
|
|
case CmpInst::FCMP_UEQ:
|
|
|
|
default:
|
|
|
|
// AL is our "false" for now. The other two need more compares.
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::AL;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_EQ:
|
|
|
|
case CmpInst::FCMP_OEQ:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::EQ;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_SGT:
|
|
|
|
case CmpInst::FCMP_OGT:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::GT;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_SGE:
|
|
|
|
case CmpInst::FCMP_OGE:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::GE;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_UGT:
|
|
|
|
case CmpInst::FCMP_UGT:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::HI;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::FCMP_OLT:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::MI;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_ULE:
|
|
|
|
case CmpInst::FCMP_OLE:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::LS;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::FCMP_ORD:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::VC;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::FCMP_UNO:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::VS;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::FCMP_UGE:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::PL;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_SLT:
|
|
|
|
case CmpInst::FCMP_ULT:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::LT;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_SLE:
|
|
|
|
case CmpInst::FCMP_ULE:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::LE;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::FCMP_UNE:
|
|
|
|
case CmpInst::ICMP_NE:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::NE;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_UGE:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::HS;
|
2014-03-29 18:18:08 +08:00
|
|
|
case CmpInst::ICMP_ULT:
|
2014-05-24 20:50:23 +08:00
|
|
|
return AArch64CC::LO;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-09-18 02:05:34 +08:00
|
|
|
/// \brief Check if the comparison against zero and the following branch can be
|
|
|
|
/// folded into a single instruction (CBZ or CBNZ).
|
2014-09-18 16:07:40 +08:00
|
|
|
static bool canFoldZeroIntoBranch(const CmpInst *CI) {
|
2014-09-18 02:05:34 +08:00
|
|
|
CmpInst::Predicate Predicate = CI->getPredicate();
|
|
|
|
if ((Predicate != CmpInst::ICMP_EQ) && (Predicate != CmpInst::ICMP_NE))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Type *Ty = CI->getOperand(0)->getType();
|
|
|
|
if (!Ty->isIntegerTy())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned BW = cast<IntegerType>(Ty)->getBitWidth();
|
|
|
|
if (BW != 1 && BW != 8 && BW != 16 && BW != 32 && BW != 64)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(CI->getOperand(0)))
|
|
|
|
if (C->isNullValue())
|
|
|
|
return true;
|
|
|
|
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(CI->getOperand(1)))
|
|
|
|
if (C->isNullValue())
|
|
|
|
return true;
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectBranch(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
const BranchInst *BI = cast<BranchInst>(I);
|
2014-09-04 01:58:10 +08:00
|
|
|
if (BI->isUnconditional()) {
|
|
|
|
MachineBasicBlock *MSucc = FuncInfo.MBBMap[BI->getSuccessor(0)];
|
2014-09-04 04:56:52 +08:00
|
|
|
fastEmitBranch(MSucc, BI->getDebugLoc());
|
2014-09-04 01:58:10 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
MachineBasicBlock *TBB = FuncInfo.MBBMap[BI->getSuccessor(0)];
|
|
|
|
MachineBasicBlock *FBB = FuncInfo.MBBMap[BI->getSuccessor(1)];
|
|
|
|
|
2014-07-31 06:04:34 +08:00
|
|
|
AArch64CC::CondCode CC = AArch64CC::NE;
|
2014-03-29 18:18:08 +08:00
|
|
|
if (const CmpInst *CI = dyn_cast<CmpInst>(BI->getCondition())) {
|
2014-09-18 01:46:47 +08:00
|
|
|
if (CI->hasOneUse() && isValueAvailable(CI)) {
|
|
|
|
// Try to optimize or fold the cmp.
|
|
|
|
CmpInst::Predicate Predicate = optimizeCmpPredicate(CI);
|
|
|
|
switch (Predicate) {
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
case CmpInst::FCMP_FALSE:
|
|
|
|
fastEmitBranch(FBB, DbgLoc);
|
|
|
|
return true;
|
|
|
|
case CmpInst::FCMP_TRUE:
|
|
|
|
fastEmitBranch(TBB, DbgLoc);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Try to take advantage of fallthrough opportunities.
|
|
|
|
if (FuncInfo.MBB->isLayoutSuccessor(TBB)) {
|
|
|
|
std::swap(TBB, FBB);
|
|
|
|
Predicate = CmpInst::getInversePredicate(Predicate);
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-09-18 02:05:34 +08:00
|
|
|
// Try to optimize comparisons against zero.
|
2014-09-18 16:07:40 +08:00
|
|
|
if (canFoldZeroIntoBranch(CI)) {
|
2014-09-18 02:05:34 +08:00
|
|
|
const Value *LHS = CI->getOperand(0);
|
|
|
|
const Value *RHS = CI->getOperand(1);
|
|
|
|
|
|
|
|
// Canonicalize zero values to the RHS.
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(LHS))
|
|
|
|
if (C->isNullValue())
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
2014-09-18 16:07:40 +08:00
|
|
|
static const unsigned OpcTable[2][2] = {
|
|
|
|
{AArch64::CBZW, AArch64::CBZX }, {AArch64::CBNZW, AArch64::CBNZX}
|
2014-09-18 02:05:34 +08:00
|
|
|
};
|
|
|
|
bool IsCmpNE = Predicate == CmpInst::ICMP_NE;
|
|
|
|
bool Is64Bit = LHS->getType()->isIntegerTy(64);
|
2014-09-18 16:07:40 +08:00
|
|
|
unsigned Opc = OpcTable[IsCmpNE][Is64Bit];
|
2014-09-18 02:05:34 +08:00
|
|
|
|
|
|
|
unsigned SrcReg = getRegForValue(LHS);
|
|
|
|
if (!SrcReg)
|
|
|
|
return false;
|
|
|
|
bool SrcIsKill = hasTrivialKill(LHS);
|
|
|
|
|
|
|
|
// Emit the combined compare and branch instruction.
|
2014-09-18 16:07:40 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(Opc))
|
|
|
|
.addReg(SrcReg, getKillRegState(SrcIsKill))
|
|
|
|
.addMBB(TBB);
|
2014-09-18 02:05:34 +08:00
|
|
|
|
|
|
|
// Obtain the branch weight and add the TrueBB to the successor list.
|
|
|
|
uint32_t BranchWeight = 0;
|
|
|
|
if (FuncInfo.BPI)
|
|
|
|
BranchWeight = FuncInfo.BPI->getEdgeWeight(BI->getParent(),
|
|
|
|
TBB->getBasicBlock());
|
|
|
|
FuncInfo.MBB->addSuccessor(TBB, BranchWeight);
|
|
|
|
|
|
|
|
fastEmitBranch(FBB, DbgLoc);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Emit the cmp.
|
2014-08-20 06:29:55 +08:00
|
|
|
if (!emitCmp(CI->getOperand(0), CI->getOperand(1), CI->isUnsigned()))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-18 01:46:47 +08:00
|
|
|
// FCMP_UEQ and FCMP_ONE cannot be checked with a single branch
|
|
|
|
// instruction.
|
|
|
|
CC = getCompareCC(Predicate);
|
|
|
|
AArch64CC::CondCode ExtraCC = AArch64CC::AL;
|
|
|
|
switch (Predicate) {
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
case CmpInst::FCMP_UEQ:
|
|
|
|
ExtraCC = AArch64CC::EQ;
|
|
|
|
CC = AArch64CC::VS;
|
|
|
|
break;
|
|
|
|
case CmpInst::FCMP_ONE:
|
|
|
|
ExtraCC = AArch64CC::MI;
|
|
|
|
CC = AArch64CC::GT;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
assert((CC != AArch64CC::AL) && "Unexpected condition code.");
|
|
|
|
|
|
|
|
// Emit the extra branch for FCMP_UEQ and FCMP_ONE.
|
|
|
|
if (ExtraCC != AArch64CC::AL) {
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::Bcc))
|
|
|
|
.addImm(ExtraCC)
|
|
|
|
.addMBB(TBB);
|
|
|
|
}
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Emit the branch.
|
2014-05-24 20:50:23 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::Bcc))
|
2014-03-29 18:18:08 +08:00
|
|
|
.addImm(CC)
|
|
|
|
.addMBB(TBB);
|
2014-08-02 02:39:24 +08:00
|
|
|
|
|
|
|
// Obtain the branch weight and add the TrueBB to the successor list.
|
|
|
|
uint32_t BranchWeight = 0;
|
|
|
|
if (FuncInfo.BPI)
|
|
|
|
BranchWeight = FuncInfo.BPI->getEdgeWeight(BI->getParent(),
|
|
|
|
TBB->getBasicBlock());
|
|
|
|
FuncInfo.MBB->addSuccessor(TBB, BranchWeight);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
fastEmitBranch(FBB, DbgLoc);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
} else if (TruncInst *TI = dyn_cast<TruncInst>(BI->getCondition())) {
|
|
|
|
MVT SrcVT;
|
2014-09-18 01:46:47 +08:00
|
|
|
if (TI->hasOneUse() && isValueAvailable(TI) &&
|
|
|
|
isTypeSupported(TI->getOperand(0)->getType(), SrcVT)) {
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned CondReg = getRegForValue(TI->getOperand(0));
|
2014-08-22 02:02:25 +08:00
|
|
|
if (!CondReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
2014-08-22 02:02:25 +08:00
|
|
|
bool CondIsKill = hasTrivialKill(TI->getOperand(0));
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Issue an extract_subreg to get the lower 32-bits.
|
2014-08-22 02:02:25 +08:00
|
|
|
if (SrcVT == MVT::i64) {
|
2014-09-04 04:56:59 +08:00
|
|
|
CondReg = fastEmitInst_extractsubreg(MVT::i32, CondReg, CondIsKill,
|
2014-05-24 20:50:23 +08:00
|
|
|
AArch64::sub_32);
|
2014-08-22 02:02:25 +08:00
|
|
|
CondIsKill = true;
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-09-04 09:29:18 +08:00
|
|
|
unsigned ANDReg = emitAnd_ri(MVT::i32, CondReg, CondIsKill, 1);
|
2014-08-22 02:02:25 +08:00
|
|
|
assert(ANDReg && "Unexpected AND instruction emission failure.");
|
2014-08-20 06:29:55 +08:00
|
|
|
emitICmp_ri(MVT::i32, ANDReg, /*IsKill=*/true, 0);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
if (FuncInfo.MBB->isLayoutSuccessor(TBB)) {
|
|
|
|
std::swap(TBB, FBB);
|
2014-05-24 20:50:23 +08:00
|
|
|
CC = AArch64CC::EQ;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
2014-05-24 20:50:23 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::Bcc))
|
2014-03-29 18:18:08 +08:00
|
|
|
.addImm(CC)
|
|
|
|
.addMBB(TBB);
|
2014-08-02 02:39:24 +08:00
|
|
|
|
|
|
|
// Obtain the branch weight and add the TrueBB to the successor list.
|
|
|
|
uint32_t BranchWeight = 0;
|
|
|
|
if (FuncInfo.BPI)
|
|
|
|
BranchWeight = FuncInfo.BPI->getEdgeWeight(BI->getParent(),
|
|
|
|
TBB->getBasicBlock());
|
|
|
|
FuncInfo.MBB->addSuccessor(TBB, BranchWeight);
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
fastEmitBranch(FBB, DbgLoc);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
2014-09-18 01:46:47 +08:00
|
|
|
} else if (const auto *CI = dyn_cast<ConstantInt>(BI->getCondition())) {
|
2014-03-29 18:18:08 +08:00
|
|
|
uint64_t Imm = CI->getZExtValue();
|
|
|
|
MachineBasicBlock *Target = (Imm == 0) ? FBB : TBB;
|
2014-05-24 20:50:23 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::B))
|
2014-03-29 18:18:08 +08:00
|
|
|
.addMBB(Target);
|
2014-08-02 02:39:24 +08:00
|
|
|
|
|
|
|
// Obtain the branch weight and add the target to the successor list.
|
|
|
|
uint32_t BranchWeight = 0;
|
|
|
|
if (FuncInfo.BPI)
|
|
|
|
BranchWeight = FuncInfo.BPI->getEdgeWeight(BI->getParent(),
|
|
|
|
Target->getBasicBlock());
|
|
|
|
FuncInfo.MBB->addSuccessor(Target, BranchWeight);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
2014-07-31 06:04:34 +08:00
|
|
|
} else if (foldXALUIntrinsic(CC, I, BI->getCondition())) {
|
|
|
|
// Fake request the condition, otherwise the intrinsic might be completely
|
|
|
|
// optimized away.
|
|
|
|
unsigned CondReg = getRegForValue(BI->getCondition());
|
|
|
|
if (!CondReg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Emit the branch.
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::Bcc))
|
|
|
|
.addImm(CC)
|
|
|
|
.addMBB(TBB);
|
2014-08-02 02:39:24 +08:00
|
|
|
|
|
|
|
// Obtain the branch weight and add the TrueBB to the successor list.
|
|
|
|
uint32_t BranchWeight = 0;
|
|
|
|
if (FuncInfo.BPI)
|
|
|
|
BranchWeight = FuncInfo.BPI->getEdgeWeight(BI->getParent(),
|
|
|
|
TBB->getBasicBlock());
|
|
|
|
FuncInfo.MBB->addSuccessor(TBB, BranchWeight);
|
2014-07-31 06:04:34 +08:00
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
fastEmitBranch(FBB, DbgLoc);
|
2014-07-31 06:04:34 +08:00
|
|
|
return true;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned CondReg = getRegForValue(BI->getCondition());
|
|
|
|
if (CondReg == 0)
|
|
|
|
return false;
|
2014-08-20 06:29:55 +08:00
|
|
|
bool CondRegIsKill = hasTrivialKill(BI->getCondition());
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// We've been divorced from our compare! Our block was split, and
|
|
|
|
// now our compare lives in a predecessor block. We musn't
|
|
|
|
// re-compare here, as the children of the compare aren't guaranteed
|
|
|
|
// live across the block boundary (we *could* check for this).
|
|
|
|
// Regardless, the compare has been done in the predecessor block,
|
|
|
|
// and it left a value for us in a virtual register. Ergo, we test
|
|
|
|
// the one-bit value left in the virtual register.
|
2014-08-20 06:29:55 +08:00
|
|
|
emitICmp_ri(MVT::i32, CondReg, CondRegIsKill, 0);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
if (FuncInfo.MBB->isLayoutSuccessor(TBB)) {
|
|
|
|
std::swap(TBB, FBB);
|
2014-05-24 20:50:23 +08:00
|
|
|
CC = AArch64CC::EQ;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::Bcc))
|
2014-03-29 18:18:08 +08:00
|
|
|
.addImm(CC)
|
|
|
|
.addMBB(TBB);
|
2014-08-02 02:39:24 +08:00
|
|
|
|
|
|
|
// Obtain the branch weight and add the TrueBB to the successor list.
|
|
|
|
uint32_t BranchWeight = 0;
|
|
|
|
if (FuncInfo.BPI)
|
|
|
|
BranchWeight = FuncInfo.BPI->getEdgeWeight(BI->getParent(),
|
|
|
|
TBB->getBasicBlock());
|
|
|
|
FuncInfo.MBB->addSuccessor(TBB, BranchWeight);
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
fastEmitBranch(FBB, DbgLoc);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectIndirectBr(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
const IndirectBrInst *BI = cast<IndirectBrInst>(I);
|
|
|
|
unsigned AddrReg = getRegForValue(BI->getOperand(0));
|
|
|
|
if (AddrReg == 0)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Emit the indirect branch.
|
2014-08-22 04:57:57 +08:00
|
|
|
const MCInstrDesc &II = TII.get(AArch64::BR);
|
|
|
|
AddrReg = constrainOperandRegClass(II, AddrReg, II.getNumDefs());
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, II).addReg(AddrReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Make sure the CFG is up-to-date.
|
|
|
|
for (unsigned i = 0, e = BI->getNumSuccessors(); i != e; ++i)
|
|
|
|
FuncInfo.MBB->addSuccessor(FuncInfo.MBBMap[BI->getSuccessor(i)]);
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectCmp(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
const CmpInst *CI = cast<CmpInst>(I);
|
|
|
|
|
2014-09-16 04:47:16 +08:00
|
|
|
// Try to optimize or fold the cmp.
|
|
|
|
CmpInst::Predicate Predicate = optimizeCmpPredicate(CI);
|
|
|
|
unsigned ResultReg = 0;
|
|
|
|
switch (Predicate) {
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
case CmpInst::FCMP_FALSE:
|
|
|
|
ResultReg = createResultReg(&AArch64::GPR32RegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(TargetOpcode::COPY), ResultReg)
|
|
|
|
.addReg(AArch64::WZR, getKillRegState(true));
|
|
|
|
break;
|
|
|
|
case CmpInst::FCMP_TRUE:
|
|
|
|
ResultReg = fastEmit_i(MVT::i32, MVT::i32, ISD::Constant, 1);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ResultReg) {
|
|
|
|
updateValueMap(I, ResultReg);
|
|
|
|
return true;
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Emit the cmp.
|
2014-08-20 06:29:55 +08:00
|
|
|
if (!emitCmp(CI->getOperand(0), CI->getOperand(1), CI->isUnsigned()))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-16 04:47:16 +08:00
|
|
|
ResultReg = createResultReg(&AArch64::GPR32RegClass);
|
|
|
|
|
|
|
|
// FCMP_UEQ and FCMP_ONE cannot be checked with a single instruction. These
|
|
|
|
// condition codes are inverted, because they are used by CSINC.
|
|
|
|
static unsigned CondCodeTable[2][2] = {
|
|
|
|
{ AArch64CC::NE, AArch64CC::VC },
|
|
|
|
{ AArch64CC::PL, AArch64CC::LE }
|
|
|
|
};
|
|
|
|
unsigned *CondCodes = nullptr;
|
|
|
|
switch (Predicate) {
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
case CmpInst::FCMP_UEQ:
|
|
|
|
CondCodes = &CondCodeTable[0][0];
|
|
|
|
break;
|
|
|
|
case CmpInst::FCMP_ONE:
|
|
|
|
CondCodes = &CondCodeTable[1][0];
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (CondCodes) {
|
|
|
|
unsigned TmpReg1 = createResultReg(&AArch64::GPR32RegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::CSINCWr),
|
|
|
|
TmpReg1)
|
|
|
|
.addReg(AArch64::WZR, getKillRegState(true))
|
|
|
|
.addReg(AArch64::WZR, getKillRegState(true))
|
|
|
|
.addImm(CondCodes[0]);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::CSINCWr),
|
|
|
|
ResultReg)
|
|
|
|
.addReg(TmpReg1, getKillRegState(true))
|
|
|
|
.addReg(AArch64::WZR, getKillRegState(true))
|
|
|
|
.addImm(CondCodes[1]);
|
|
|
|
|
|
|
|
updateValueMap(I, ResultReg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Now set a register based on the comparison.
|
2014-09-16 04:47:16 +08:00
|
|
|
AArch64CC::CondCode CC = getCompareCC(Predicate);
|
|
|
|
assert((CC != AArch64CC::AL) && "Unexpected condition code.");
|
2014-05-24 20:50:23 +08:00
|
|
|
AArch64CC::CondCode invertedCC = getInvertedCondCode(CC);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::CSINCWr),
|
2014-03-29 18:18:08 +08:00
|
|
|
ResultReg)
|
2014-09-16 04:47:16 +08:00
|
|
|
.addReg(AArch64::WZR, getKillRegState(true))
|
|
|
|
.addReg(AArch64::WZR, getKillRegState(true))
|
2014-03-29 18:18:08 +08:00
|
|
|
.addImm(invertedCC);
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectSelect(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
const SelectInst *SI = cast<SelectInst>(I);
|
|
|
|
|
|
|
|
EVT DestEVT = TLI.getValueType(SI->getType(), true);
|
|
|
|
if (!DestEVT.isSimple())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MVT DestVT = DestEVT.getSimpleVT();
|
|
|
|
if (DestVT != MVT::i32 && DestVT != MVT::i64 && DestVT != MVT::f32 &&
|
|
|
|
DestVT != MVT::f64)
|
|
|
|
return false;
|
|
|
|
|
2014-07-31 06:04:37 +08:00
|
|
|
unsigned SelectOpc;
|
2014-08-22 04:57:57 +08:00
|
|
|
const TargetRegisterClass *RC = nullptr;
|
2014-07-31 06:04:37 +08:00
|
|
|
switch (DestVT.SimpleTy) {
|
|
|
|
default: return false;
|
2014-08-22 04:57:57 +08:00
|
|
|
case MVT::i32:
|
|
|
|
SelectOpc = AArch64::CSELWr; RC = &AArch64::GPR32RegClass; break;
|
|
|
|
case MVT::i64:
|
|
|
|
SelectOpc = AArch64::CSELXr; RC = &AArch64::GPR64RegClass; break;
|
|
|
|
case MVT::f32:
|
|
|
|
SelectOpc = AArch64::FCSELSrrr; RC = &AArch64::FPR32RegClass; break;
|
|
|
|
case MVT::f64:
|
|
|
|
SelectOpc = AArch64::FCSELDrrr; RC = &AArch64::FPR64RegClass; break;
|
2014-07-31 06:04:37 +08:00
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-07-31 06:04:37 +08:00
|
|
|
const Value *Cond = SI->getCondition();
|
|
|
|
bool NeedTest = true;
|
|
|
|
AArch64CC::CondCode CC = AArch64CC::NE;
|
|
|
|
if (foldXALUIntrinsic(CC, I, Cond))
|
|
|
|
NeedTest = false;
|
2014-05-04 01:27:06 +08:00
|
|
|
|
2014-07-31 06:04:37 +08:00
|
|
|
unsigned CondReg = getRegForValue(Cond);
|
|
|
|
if (!CondReg)
|
|
|
|
return false;
|
|
|
|
bool CondIsKill = hasTrivialKill(Cond);
|
|
|
|
|
|
|
|
if (NeedTest) {
|
2014-09-04 09:29:18 +08:00
|
|
|
unsigned ANDReg = emitAnd_ri(MVT::i32, CondReg, CondIsKill, 1);
|
2014-08-22 02:02:25 +08:00
|
|
|
assert(ANDReg && "Unexpected AND instruction emission failure.");
|
|
|
|
emitICmp_ri(MVT::i32, ANDReg, /*IsKill=*/true, 0);
|
2014-07-31 06:04:37 +08:00
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-07-31 06:04:37 +08:00
|
|
|
unsigned TrueReg = getRegForValue(SI->getTrueValue());
|
|
|
|
bool TrueIsKill = hasTrivialKill(SI->getTrueValue());
|
|
|
|
|
|
|
|
unsigned FalseReg = getRegForValue(SI->getFalseValue());
|
|
|
|
bool FalseIsKill = hasTrivialKill(SI->getFalseValue());
|
|
|
|
|
|
|
|
if (!TrueReg || !FalseReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned ResultReg = fastEmitInst_rri(SelectOpc, RC, TrueReg, TrueIsKill,
|
2014-08-22 04:57:57 +08:00
|
|
|
FalseReg, FalseIsKill, CC);
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectFPExt(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
Value *V = I->getOperand(0);
|
|
|
|
if (!I->getType()->isDoubleTy() || !V->getType()->isFloatTy())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned Op = getRegForValue(V);
|
|
|
|
if (Op == 0)
|
|
|
|
return false;
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
unsigned ResultReg = createResultReg(&AArch64::FPR64RegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::FCVTDSr),
|
2014-03-29 18:18:08 +08:00
|
|
|
ResultReg).addReg(Op);
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectFPTrunc(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
Value *V = I->getOperand(0);
|
|
|
|
if (!I->getType()->isFloatTy() || !V->getType()->isDoubleTy())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned Op = getRegForValue(V);
|
|
|
|
if (Op == 0)
|
|
|
|
return false;
|
|
|
|
|
2014-05-24 20:50:23 +08:00
|
|
|
unsigned ResultReg = createResultReg(&AArch64::FPR32RegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::FCVTSDr),
|
2014-03-29 18:18:08 +08:00
|
|
|
ResultReg).addReg(Op);
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// FPToUI and FPToSI
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectFPToInt(const Instruction *I, bool Signed) {
|
2014-03-29 18:18:08 +08:00
|
|
|
MVT DestVT;
|
|
|
|
if (!isTypeLegal(I->getType(), DestVT) || DestVT.isVector())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned SrcReg = getRegForValue(I->getOperand(0));
|
|
|
|
if (SrcReg == 0)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
EVT SrcVT = TLI.getValueType(I->getOperand(0)->getType(), true);
|
2014-04-30 23:29:57 +08:00
|
|
|
if (SrcVT == MVT::f128)
|
|
|
|
return false;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
unsigned Opc;
|
|
|
|
if (SrcVT == MVT::f64) {
|
|
|
|
if (Signed)
|
2014-05-24 20:50:23 +08:00
|
|
|
Opc = (DestVT == MVT::i32) ? AArch64::FCVTZSUWDr : AArch64::FCVTZSUXDr;
|
2014-03-29 18:18:08 +08:00
|
|
|
else
|
2014-05-24 20:50:23 +08:00
|
|
|
Opc = (DestVT == MVT::i32) ? AArch64::FCVTZUUWDr : AArch64::FCVTZUUXDr;
|
2014-03-29 18:18:08 +08:00
|
|
|
} else {
|
|
|
|
if (Signed)
|
2014-05-24 20:50:23 +08:00
|
|
|
Opc = (DestVT == MVT::i32) ? AArch64::FCVTZSUWSr : AArch64::FCVTZSUXSr;
|
2014-03-29 18:18:08 +08:00
|
|
|
else
|
2014-05-24 20:50:23 +08:00
|
|
|
Opc = (DestVT == MVT::i32) ? AArch64::FCVTZUUWSr : AArch64::FCVTZUUXSr;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
2014-04-15 21:59:53 +08:00
|
|
|
unsigned ResultReg = createResultReg(
|
2014-05-24 20:50:23 +08:00
|
|
|
DestVT == MVT::i32 ? &AArch64::GPR32RegClass : &AArch64::GPR64RegClass);
|
2014-03-29 18:18:08 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(Opc), ResultReg)
|
|
|
|
.addReg(SrcReg);
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectIntToFP(const Instruction *I, bool Signed) {
|
2014-03-29 18:18:08 +08:00
|
|
|
MVT DestVT;
|
|
|
|
if (!isTypeLegal(I->getType(), DestVT) || DestVT.isVector())
|
|
|
|
return false;
|
2014-04-30 23:29:57 +08:00
|
|
|
assert ((DestVT == MVT::f32 || DestVT == MVT::f64) &&
|
|
|
|
"Unexpected value type.");
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
unsigned SrcReg = getRegForValue(I->getOperand(0));
|
2014-08-22 04:57:57 +08:00
|
|
|
if (!SrcReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
2014-08-22 04:57:57 +08:00
|
|
|
bool SrcIsKill = hasTrivialKill(I->getOperand(0));
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
EVT SrcVT = TLI.getValueType(I->getOperand(0)->getType(), true);
|
|
|
|
|
|
|
|
// Handle sign-extension.
|
|
|
|
if (SrcVT == MVT::i16 || SrcVT == MVT::i8 || SrcVT == MVT::i1) {
|
|
|
|
SrcReg =
|
2014-09-16 07:20:17 +08:00
|
|
|
emitIntExt(SrcVT.getSimpleVT(), SrcReg, MVT::i32, /*isZExt*/ !Signed);
|
2014-08-22 04:57:57 +08:00
|
|
|
if (!SrcReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
2014-08-22 04:57:57 +08:00
|
|
|
SrcIsKill = true;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned Opc;
|
|
|
|
if (SrcVT == MVT::i64) {
|
|
|
|
if (Signed)
|
2014-05-24 20:50:23 +08:00
|
|
|
Opc = (DestVT == MVT::f32) ? AArch64::SCVTFUXSri : AArch64::SCVTFUXDri;
|
2014-03-29 18:18:08 +08:00
|
|
|
else
|
2014-05-24 20:50:23 +08:00
|
|
|
Opc = (DestVT == MVT::f32) ? AArch64::UCVTFUXSri : AArch64::UCVTFUXDri;
|
2014-03-29 18:18:08 +08:00
|
|
|
} else {
|
|
|
|
if (Signed)
|
2014-05-24 20:50:23 +08:00
|
|
|
Opc = (DestVT == MVT::f32) ? AArch64::SCVTFUWSri : AArch64::SCVTFUWDri;
|
2014-03-29 18:18:08 +08:00
|
|
|
else
|
2014-05-24 20:50:23 +08:00
|
|
|
Opc = (DestVT == MVT::f32) ? AArch64::UCVTFUWSri : AArch64::UCVTFUWDri;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned ResultReg = fastEmitInst_r(Opc, TLI.getRegClassFor(DestVT), SrcReg,
|
2014-08-22 04:57:57 +08:00
|
|
|
SrcIsKill);
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
bool AArch64FastISel::fastLowerArguments() {
|
2014-08-05 13:43:48 +08:00
|
|
|
if (!FuncInfo.CanLowerReturn)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const Function *F = FuncInfo.Fn;
|
|
|
|
if (F->isVarArg())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
CallingConv::ID CC = F->getCallingConv();
|
|
|
|
if (CC != CallingConv::C)
|
|
|
|
return false;
|
|
|
|
|
2014-09-16 08:25:30 +08:00
|
|
|
// Only handle simple cases of up to 8 GPR and FPR each.
|
2014-08-05 13:43:48 +08:00
|
|
|
unsigned GPRCnt = 0;
|
|
|
|
unsigned FPRCnt = 0;
|
|
|
|
unsigned Idx = 0;
|
|
|
|
for (auto const &Arg : F->args()) {
|
|
|
|
// The first argument is at index 1.
|
|
|
|
++Idx;
|
|
|
|
if (F->getAttributes().hasAttribute(Idx, Attribute::ByVal) ||
|
|
|
|
F->getAttributes().hasAttribute(Idx, Attribute::InReg) ||
|
|
|
|
F->getAttributes().hasAttribute(Idx, Attribute::StructRet) ||
|
|
|
|
F->getAttributes().hasAttribute(Idx, Attribute::Nest))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Type *ArgTy = Arg.getType();
|
2014-09-16 08:25:30 +08:00
|
|
|
if (ArgTy->isStructTy() || ArgTy->isArrayTy())
|
2014-08-05 13:43:48 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
EVT ArgVT = TLI.getValueType(ArgTy);
|
2014-09-16 08:25:30 +08:00
|
|
|
if (!ArgVT.isSimple())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MVT VT = ArgVT.getSimpleVT().SimpleTy;
|
|
|
|
if (VT.isFloatingPoint() && !Subtarget->hasFPARMv8())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (VT.isVector() &&
|
|
|
|
(!Subtarget->hasNEON() || !Subtarget->isLittleEndian()))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (VT >= MVT::i1 && VT <= MVT::i64)
|
2014-08-05 13:43:48 +08:00
|
|
|
++GPRCnt;
|
2014-09-16 08:25:30 +08:00
|
|
|
else if ((VT >= MVT::f16 && VT <= MVT::f64) || VT.is64BitVector() ||
|
|
|
|
VT.is128BitVector())
|
2014-08-05 13:43:48 +08:00
|
|
|
++FPRCnt;
|
2014-09-16 08:25:30 +08:00
|
|
|
else
|
|
|
|
return false;
|
2014-08-05 13:43:48 +08:00
|
|
|
|
|
|
|
if (GPRCnt > 8 || FPRCnt > 8)
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2014-09-16 08:25:30 +08:00
|
|
|
static const MCPhysReg Registers[6][8] = {
|
2014-08-05 13:43:48 +08:00
|
|
|
{ AArch64::W0, AArch64::W1, AArch64::W2, AArch64::W3, AArch64::W4,
|
|
|
|
AArch64::W5, AArch64::W6, AArch64::W7 },
|
|
|
|
{ AArch64::X0, AArch64::X1, AArch64::X2, AArch64::X3, AArch64::X4,
|
|
|
|
AArch64::X5, AArch64::X6, AArch64::X7 },
|
|
|
|
{ AArch64::H0, AArch64::H1, AArch64::H2, AArch64::H3, AArch64::H4,
|
|
|
|
AArch64::H5, AArch64::H6, AArch64::H7 },
|
|
|
|
{ AArch64::S0, AArch64::S1, AArch64::S2, AArch64::S3, AArch64::S4,
|
|
|
|
AArch64::S5, AArch64::S6, AArch64::S7 },
|
|
|
|
{ AArch64::D0, AArch64::D1, AArch64::D2, AArch64::D3, AArch64::D4,
|
2014-09-16 08:25:30 +08:00
|
|
|
AArch64::D5, AArch64::D6, AArch64::D7 },
|
|
|
|
{ AArch64::Q0, AArch64::Q1, AArch64::Q2, AArch64::Q3, AArch64::Q4,
|
|
|
|
AArch64::Q5, AArch64::Q6, AArch64::Q7 }
|
2014-08-05 13:43:48 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
unsigned GPRIdx = 0;
|
|
|
|
unsigned FPRIdx = 0;
|
|
|
|
for (auto const &Arg : F->args()) {
|
|
|
|
MVT VT = TLI.getSimpleValueType(Arg.getType());
|
|
|
|
unsigned SrcReg;
|
2014-09-16 08:25:30 +08:00
|
|
|
const TargetRegisterClass *RC;
|
|
|
|
if (VT >= MVT::i1 && VT <= MVT::i32) {
|
|
|
|
SrcReg = Registers[0][GPRIdx++];
|
|
|
|
RC = &AArch64::GPR32RegClass;
|
|
|
|
VT = MVT::i32;
|
|
|
|
} else if (VT == MVT::i64) {
|
|
|
|
SrcReg = Registers[1][GPRIdx++];
|
|
|
|
RC = &AArch64::GPR64RegClass;
|
|
|
|
} else if (VT == MVT::f16) {
|
|
|
|
SrcReg = Registers[2][FPRIdx++];
|
|
|
|
RC = &AArch64::FPR16RegClass;
|
|
|
|
} else if (VT == MVT::f32) {
|
|
|
|
SrcReg = Registers[3][FPRIdx++];
|
|
|
|
RC = &AArch64::FPR32RegClass;
|
|
|
|
} else if ((VT == MVT::f64) || VT.is64BitVector()) {
|
|
|
|
SrcReg = Registers[4][FPRIdx++];
|
|
|
|
RC = &AArch64::FPR64RegClass;
|
|
|
|
} else if (VT.is128BitVector()) {
|
|
|
|
SrcReg = Registers[5][FPRIdx++];
|
|
|
|
RC = &AArch64::FPR128RegClass;
|
|
|
|
} else
|
|
|
|
llvm_unreachable("Unexpected value type.");
|
2014-08-05 13:43:48 +08:00
|
|
|
|
|
|
|
unsigned DstReg = FuncInfo.MF->addLiveIn(SrcReg, RC);
|
|
|
|
// FIXME: Unfortunately it's necessary to emit a copy from the livein copy.
|
|
|
|
// Without this, EmitLiveInCopies may eliminate the livein if its only
|
|
|
|
// use is a bitcast (which isn't turned into an instruction).
|
|
|
|
unsigned ResultReg = createResultReg(RC);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(TargetOpcode::COPY), ResultReg)
|
2014-08-22 04:57:57 +08:00
|
|
|
.addReg(DstReg, getKillRegState(true));
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(&Arg, ResultReg);
|
2014-08-05 13:43:48 +08:00
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::processCallArgs(CallLoweringInfo &CLI,
|
2014-07-23 07:14:58 +08:00
|
|
|
SmallVectorImpl<MVT> &OutVTs,
|
|
|
|
unsigned &NumBytes) {
|
|
|
|
CallingConv::ID CC = CLI.CallConv;
|
2014-03-29 18:18:08 +08:00
|
|
|
SmallVector<CCValAssign, 16> ArgLocs;
|
2014-08-07 02:45:26 +08:00
|
|
|
CCState CCInfo(CC, false, *FuncInfo.MF, ArgLocs, *Context);
|
2014-07-23 07:14:58 +08:00
|
|
|
CCInfo.AnalyzeCallOperands(OutVTs, CLI.OutFlags, CCAssignFnForCall(CC));
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Get a count of how many bytes are to be pushed on the stack.
|
|
|
|
NumBytes = CCInfo.getNextStackOffset();
|
|
|
|
|
|
|
|
// Issue CALLSEQ_START
|
|
|
|
unsigned AdjStackDown = TII.getCallFrameSetupOpcode();
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AdjStackDown))
|
2014-07-23 07:14:58 +08:00
|
|
|
.addImm(NumBytes);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Process the args.
|
|
|
|
for (unsigned i = 0, e = ArgLocs.size(); i != e; ++i) {
|
|
|
|
CCValAssign &VA = ArgLocs[i];
|
2014-07-23 07:14:58 +08:00
|
|
|
const Value *ArgVal = CLI.OutVals[VA.getValNo()];
|
|
|
|
MVT ArgVT = OutVTs[VA.getValNo()];
|
|
|
|
|
|
|
|
unsigned ArgReg = getRegForValue(ArgVal);
|
|
|
|
if (!ArgReg)
|
|
|
|
return false;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Handle arg promotion: SExt, ZExt, AExt.
|
|
|
|
switch (VA.getLocInfo()) {
|
|
|
|
case CCValAssign::Full:
|
|
|
|
break;
|
|
|
|
case CCValAssign::SExt: {
|
|
|
|
MVT DestVT = VA.getLocVT();
|
|
|
|
MVT SrcVT = ArgVT;
|
2014-09-16 07:20:17 +08:00
|
|
|
ArgReg = emitIntExt(SrcVT, ArgReg, DestVT, /*isZExt=*/false);
|
2014-07-23 07:14:58 +08:00
|
|
|
if (!ArgReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case CCValAssign::AExt:
|
|
|
|
// Intentional fall-through.
|
|
|
|
case CCValAssign::ZExt: {
|
|
|
|
MVT DestVT = VA.getLocVT();
|
|
|
|
MVT SrcVT = ArgVT;
|
2014-09-16 07:20:17 +08:00
|
|
|
ArgReg = emitIntExt(SrcVT, ArgReg, DestVT, /*isZExt=*/true);
|
2014-07-23 07:14:58 +08:00
|
|
|
if (!ArgReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unknown arg promotion!");
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now copy/store arg to correct locations.
|
|
|
|
if (VA.isRegLoc() && !VA.needsCustom()) {
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
2014-07-23 07:14:58 +08:00
|
|
|
TII.get(TargetOpcode::COPY), VA.getLocReg()).addReg(ArgReg);
|
|
|
|
CLI.OutRegs.push_back(VA.getLocReg());
|
2014-03-29 18:18:08 +08:00
|
|
|
} else if (VA.needsCustom()) {
|
|
|
|
// FIXME: Handle custom args.
|
|
|
|
return false;
|
|
|
|
} else {
|
|
|
|
assert(VA.isMemLoc() && "Assuming store on stack.");
|
|
|
|
|
2014-07-31 08:11:11 +08:00
|
|
|
// Don't emit stores for undef values.
|
|
|
|
if (isa<UndefValue>(ArgVal))
|
|
|
|
continue;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Need to store on the stack.
|
2014-06-03 21:54:53 +08:00
|
|
|
unsigned ArgSize = (ArgVT.getSizeInBits() + 7) / 8;
|
2014-05-08 20:53:50 +08:00
|
|
|
|
|
|
|
unsigned BEAlign = 0;
|
|
|
|
if (ArgSize < 8 && !Subtarget->isLittleEndian())
|
|
|
|
BEAlign = 8 - ArgSize;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
Address Addr;
|
|
|
|
Addr.setKind(Address::RegBase);
|
2014-05-24 20:50:23 +08:00
|
|
|
Addr.setReg(AArch64::SP);
|
2014-05-08 20:53:50 +08:00
|
|
|
Addr.setOffset(VA.getLocMemOffset() + BEAlign);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-08-09 01:24:10 +08:00
|
|
|
unsigned Alignment = DL.getABITypeAlignment(ArgVal->getType());
|
|
|
|
MachineMemOperand *MMO = FuncInfo.MF->getMachineMemOperand(
|
|
|
|
MachinePointerInfo::getStack(Addr.getOffset()),
|
|
|
|
MachineMemOperand::MOStore, ArgVT.getStoreSize(), Alignment);
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!emitStore(ArgVT, ArgReg, Addr, MMO))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::finishCall(CallLoweringInfo &CLI, MVT RetVT,
|
2014-07-24 04:03:13 +08:00
|
|
|
unsigned NumBytes) {
|
2014-07-23 07:14:58 +08:00
|
|
|
CallingConv::ID CC = CLI.CallConv;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Issue CALLSEQ_END
|
|
|
|
unsigned AdjStackUp = TII.getCallFrameDestroyOpcode();
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AdjStackUp))
|
2014-07-23 07:14:58 +08:00
|
|
|
.addImm(NumBytes).addImm(0);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Now the return value.
|
|
|
|
if (RetVT != MVT::isVoid) {
|
|
|
|
SmallVector<CCValAssign, 16> RVLocs;
|
2014-08-07 02:45:26 +08:00
|
|
|
CCState CCInfo(CC, false, *FuncInfo.MF, RVLocs, *Context);
|
2014-03-29 18:18:08 +08:00
|
|
|
CCInfo.AnalyzeCallResult(RetVT, CCAssignFnForCall(CC));
|
|
|
|
|
|
|
|
// Only handle a single return value.
|
|
|
|
if (RVLocs.size() != 1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Copy all of the result registers out of their specified physreg.
|
|
|
|
MVT CopyVT = RVLocs[0].getValVT();
|
|
|
|
unsigned ResultReg = createResultReg(TLI.getRegClassFor(CopyVT));
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
2014-07-23 07:14:58 +08:00
|
|
|
TII.get(TargetOpcode::COPY), ResultReg)
|
2014-08-22 04:57:57 +08:00
|
|
|
.addReg(RVLocs[0].getLocReg());
|
2014-07-23 07:14:58 +08:00
|
|
|
CLI.InRegs.push_back(RVLocs[0].getLocReg());
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
CLI.ResultReg = ResultReg;
|
|
|
|
CLI.NumResultRegs = 1;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
bool AArch64FastISel::fastLowerCall(CallLoweringInfo &CLI) {
|
2014-07-23 07:14:58 +08:00
|
|
|
CallingConv::ID CC = CLI.CallConv;
|
2014-08-14 07:23:58 +08:00
|
|
|
bool IsTailCall = CLI.IsTailCall;
|
2014-07-23 07:14:58 +08:00
|
|
|
bool IsVarArg = CLI.IsVarArg;
|
|
|
|
const Value *Callee = CLI.Callee;
|
|
|
|
const char *SymName = CLI.SymName;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-09-16 06:07:49 +08:00
|
|
|
if (!Callee && !SymName)
|
|
|
|
return false;
|
|
|
|
|
2014-08-14 07:23:58 +08:00
|
|
|
// Allow SelectionDAG isel to handle tail calls.
|
|
|
|
if (IsTailCall)
|
|
|
|
return false;
|
|
|
|
|
2014-07-31 12:10:40 +08:00
|
|
|
CodeModel::Model CM = TM.getCodeModel();
|
|
|
|
// Only support the small and large code model.
|
|
|
|
if (CM != CodeModel::Small && CM != CodeModel::Large)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// FIXME: Add large code model support for ELF.
|
|
|
|
if (CM == CodeModel::Large && !Subtarget->isTargetMachO())
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// Let SDISel handle vararg functions.
|
2014-07-23 07:14:58 +08:00
|
|
|
if (IsVarArg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
// FIXME: Only handle *simple* calls for now.
|
2014-03-29 18:18:08 +08:00
|
|
|
MVT RetVT;
|
2014-07-23 07:14:58 +08:00
|
|
|
if (CLI.RetTy->isVoidTy())
|
2014-03-29 18:18:08 +08:00
|
|
|
RetVT = MVT::isVoid;
|
2014-07-23 07:14:58 +08:00
|
|
|
else if (!isTypeLegal(CLI.RetTy, RetVT))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
for (auto Flag : CLI.OutFlags)
|
|
|
|
if (Flag.isInReg() || Flag.isSRet() || Flag.isNest() || Flag.isByVal())
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
// Set up the argument vectors.
|
|
|
|
SmallVector<MVT, 16> OutVTs;
|
|
|
|
OutVTs.reserve(CLI.OutVals.size());
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
for (auto *Val : CLI.OutVals) {
|
|
|
|
MVT VT;
|
|
|
|
if (!isTypeLegal(Val->getType(), VT) &&
|
|
|
|
!(VT == MVT::i1 || VT == MVT::i8 || VT == MVT::i16))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// We don't handle vector parameters yet.
|
2014-07-23 07:14:58 +08:00
|
|
|
if (VT.isVector() || VT.getSizeInBits() > 64)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
OutVTs.push_back(VT);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-07-31 12:10:40 +08:00
|
|
|
Address Addr;
|
2014-09-16 07:20:17 +08:00
|
|
|
if (Callee && !computeCallAddress(Callee, Addr))
|
2014-07-31 12:10:40 +08:00
|
|
|
return false;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Handle the arguments now that we've gotten them.
|
|
|
|
unsigned NumBytes;
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!processCallArgs(CLI, OutVTs, NumBytes))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// Issue the call.
|
|
|
|
MachineInstrBuilder MIB;
|
2014-07-31 12:10:40 +08:00
|
|
|
if (CM == CodeModel::Small) {
|
2014-08-30 07:48:06 +08:00
|
|
|
const MCInstrDesc &II = TII.get(Addr.getReg() ? AArch64::BLR : AArch64::BL);
|
|
|
|
MIB = BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, II);
|
2014-07-31 12:10:40 +08:00
|
|
|
if (SymName)
|
|
|
|
MIB.addExternalSymbol(SymName, 0);
|
|
|
|
else if (Addr.getGlobalValue())
|
|
|
|
MIB.addGlobalAddress(Addr.getGlobalValue(), 0, 0);
|
2014-08-30 07:48:06 +08:00
|
|
|
else if (Addr.getReg()) {
|
|
|
|
unsigned Reg = constrainOperandRegClass(II, Addr.getReg(), 0);
|
|
|
|
MIB.addReg(Reg);
|
|
|
|
} else
|
2014-07-31 12:10:40 +08:00
|
|
|
return false;
|
|
|
|
} else {
|
|
|
|
unsigned CallReg = 0;
|
|
|
|
if (SymName) {
|
|
|
|
unsigned ADRPReg = createResultReg(&AArch64::GPR64commonRegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::ADRP),
|
|
|
|
ADRPReg)
|
|
|
|
.addExternalSymbol(SymName, AArch64II::MO_GOT | AArch64II::MO_PAGE);
|
|
|
|
|
|
|
|
CallReg = createResultReg(&AArch64::GPR64RegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::LDRXui),
|
|
|
|
CallReg)
|
|
|
|
.addReg(ADRPReg)
|
|
|
|
.addExternalSymbol(SymName, AArch64II::MO_GOT | AArch64II::MO_PAGEOFF |
|
|
|
|
AArch64II::MO_NC);
|
2014-09-16 07:20:17 +08:00
|
|
|
} else if (Addr.getGlobalValue())
|
|
|
|
CallReg = materializeGV(Addr.getGlobalValue());
|
|
|
|
else if (Addr.getReg())
|
2014-07-31 12:10:40 +08:00
|
|
|
CallReg = Addr.getReg();
|
|
|
|
|
|
|
|
if (!CallReg)
|
|
|
|
return false;
|
|
|
|
|
2014-08-30 07:48:06 +08:00
|
|
|
const MCInstrDesc &II = TII.get(AArch64::BLR);
|
|
|
|
CallReg = constrainOperandRegClass(II, CallReg, 0);
|
|
|
|
MIB = BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, II).addReg(CallReg);
|
2014-07-31 12:10:40 +08:00
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Add implicit physical register uses to the call.
|
2014-07-23 07:14:58 +08:00
|
|
|
for (auto Reg : CLI.OutRegs)
|
|
|
|
MIB.addReg(Reg, RegState::Implicit);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// Add a register mask with the call-preserved registers.
|
|
|
|
// Proper defs for return values will be added by setPhysRegsDeadExcept().
|
2014-07-23 07:14:58 +08:00
|
|
|
MIB.addRegMask(TRI.getCallPreservedMask(CC));
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-07-31 12:10:40 +08:00
|
|
|
CLI.Call = MIB;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Finish off the call including any return values.
|
2014-09-16 07:20:17 +08:00
|
|
|
return finishCall(CLI, RetVT, NumBytes);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::isMemCpySmall(uint64_t Len, unsigned Alignment) {
|
2014-03-29 18:18:08 +08:00
|
|
|
if (Alignment)
|
|
|
|
return Len / Alignment <= 4;
|
|
|
|
else
|
|
|
|
return Len < 32;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::tryEmitSmallMemCpy(Address Dest, Address Src,
|
2014-05-24 20:50:23 +08:00
|
|
|
uint64_t Len, unsigned Alignment) {
|
2014-03-29 18:18:08 +08:00
|
|
|
// Make sure we don't bloat code by inlining very large memcpy's.
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!isMemCpySmall(Len, Alignment))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
int64_t UnscaledOffset = 0;
|
|
|
|
Address OrigDest = Dest;
|
|
|
|
Address OrigSrc = Src;
|
|
|
|
|
|
|
|
while (Len) {
|
|
|
|
MVT VT;
|
|
|
|
if (!Alignment || Alignment >= 8) {
|
|
|
|
if (Len >= 8)
|
|
|
|
VT = MVT::i64;
|
|
|
|
else if (Len >= 4)
|
|
|
|
VT = MVT::i32;
|
|
|
|
else if (Len >= 2)
|
|
|
|
VT = MVT::i16;
|
|
|
|
else {
|
|
|
|
VT = MVT::i8;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Bound based on alignment.
|
|
|
|
if (Len >= 4 && Alignment == 4)
|
|
|
|
VT = MVT::i32;
|
|
|
|
else if (Len >= 2 && Alignment == 2)
|
|
|
|
VT = MVT::i16;
|
|
|
|
else {
|
|
|
|
VT = MVT::i8;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool RV;
|
|
|
|
unsigned ResultReg;
|
2014-09-16 07:20:17 +08:00
|
|
|
RV = emitLoad(VT, ResultReg, Src);
|
2014-06-10 17:52:40 +08:00
|
|
|
if (!RV)
|
|
|
|
return false;
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
RV = emitStore(VT, ResultReg, Dest);
|
2014-06-10 17:52:40 +08:00
|
|
|
if (!RV)
|
|
|
|
return false;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
int64_t Size = VT.getSizeInBits() / 8;
|
|
|
|
Len -= Size;
|
|
|
|
UnscaledOffset += Size;
|
|
|
|
|
|
|
|
// We need to recompute the unscaled offset for each iteration.
|
|
|
|
Dest.setOffset(OrigDest.getOffset() + UnscaledOffset);
|
|
|
|
Src.setOffset(OrigSrc.getOffset() + UnscaledOffset);
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-07-31 06:04:34 +08:00
|
|
|
/// \brief Check if it is possible to fold the condition from the XALU intrinsic
|
|
|
|
/// into the user. The condition code will only be updated on success.
|
|
|
|
bool AArch64FastISel::foldXALUIntrinsic(AArch64CC::CondCode &CC,
|
|
|
|
const Instruction *I,
|
|
|
|
const Value *Cond) {
|
|
|
|
if (!isa<ExtractValueInst>(Cond))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const auto *EV = cast<ExtractValueInst>(Cond);
|
|
|
|
if (!isa<IntrinsicInst>(EV->getAggregateOperand()))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const auto *II = cast<IntrinsicInst>(EV->getAggregateOperand());
|
|
|
|
MVT RetVT;
|
|
|
|
const Function *Callee = II->getCalledFunction();
|
|
|
|
Type *RetTy =
|
|
|
|
cast<StructType>(Callee->getReturnType())->getTypeAtIndex(0U);
|
|
|
|
if (!isTypeLegal(RetTy, RetVT))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (RetVT != MVT::i32 && RetVT != MVT::i64)
|
|
|
|
return false;
|
|
|
|
|
2014-09-18 15:26:26 +08:00
|
|
|
const Value *LHS = II->getArgOperand(0);
|
|
|
|
const Value *RHS = II->getArgOperand(1);
|
|
|
|
|
|
|
|
// Canonicalize immediate to the RHS.
|
|
|
|
if (isa<ConstantInt>(LHS) && !isa<ConstantInt>(RHS) &&
|
|
|
|
isCommutativeIntrinsic(II))
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
|
|
|
// Simplify multiplies.
|
|
|
|
unsigned IID = II->getIntrinsicID();
|
|
|
|
switch (IID) {
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
case Intrinsic::smul_with_overflow:
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(RHS))
|
|
|
|
if (C->getValue() == 2)
|
|
|
|
IID = Intrinsic::sadd_with_overflow;
|
|
|
|
break;
|
|
|
|
case Intrinsic::umul_with_overflow:
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(RHS))
|
|
|
|
if (C->getValue() == 2)
|
|
|
|
IID = Intrinsic::uadd_with_overflow;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2014-07-31 06:04:34 +08:00
|
|
|
AArch64CC::CondCode TmpCC;
|
2014-09-18 15:26:26 +08:00
|
|
|
switch (IID) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case Intrinsic::sadd_with_overflow:
|
|
|
|
case Intrinsic::ssub_with_overflow:
|
|
|
|
TmpCC = AArch64CC::VS;
|
|
|
|
break;
|
|
|
|
case Intrinsic::uadd_with_overflow:
|
|
|
|
TmpCC = AArch64CC::HS;
|
|
|
|
break;
|
|
|
|
case Intrinsic::usub_with_overflow:
|
|
|
|
TmpCC = AArch64CC::LO;
|
|
|
|
break;
|
|
|
|
case Intrinsic::smul_with_overflow:
|
|
|
|
case Intrinsic::umul_with_overflow:
|
|
|
|
TmpCC = AArch64CC::NE;
|
|
|
|
break;
|
2014-07-31 06:04:34 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Check if both instructions are in the same basic block.
|
2014-09-18 01:46:47 +08:00
|
|
|
if (!isValueAvailable(II))
|
2014-07-31 06:04:34 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// Make sure nothing is in the way
|
|
|
|
BasicBlock::const_iterator Start = I;
|
|
|
|
BasicBlock::const_iterator End = II;
|
|
|
|
for (auto Itr = std::prev(Start); Itr != End; --Itr) {
|
|
|
|
// We only expect extractvalue instructions between the intrinsic and the
|
|
|
|
// instruction to be selected.
|
|
|
|
if (!isa<ExtractValueInst>(Itr))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Check that the extractvalue operand comes from the intrinsic.
|
|
|
|
const auto *EVI = cast<ExtractValueInst>(Itr);
|
|
|
|
if (EVI->getAggregateOperand() != II)
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
CC = TmpCC;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
bool AArch64FastISel::fastLowerIntrinsicCall(const IntrinsicInst *II) {
|
2014-03-29 18:18:08 +08:00
|
|
|
// FIXME: Handle more intrinsics.
|
2014-07-23 07:14:58 +08:00
|
|
|
switch (II->getIntrinsicID()) {
|
2014-07-26 01:47:14 +08:00
|
|
|
default: return false;
|
|
|
|
case Intrinsic::frameaddress: {
|
|
|
|
MachineFrameInfo *MFI = FuncInfo.MF->getFrameInfo();
|
|
|
|
MFI->setFrameAddressIsTaken(true);
|
|
|
|
|
|
|
|
const AArch64RegisterInfo *RegInfo =
|
2014-08-05 05:25:23 +08:00
|
|
|
static_cast<const AArch64RegisterInfo *>(
|
|
|
|
TM.getSubtargetImpl()->getRegisterInfo());
|
2014-07-26 01:47:14 +08:00
|
|
|
unsigned FramePtr = RegInfo->getFrameRegister(*(FuncInfo.MF));
|
2014-08-22 04:57:57 +08:00
|
|
|
unsigned SrcReg = MRI.createVirtualRegister(&AArch64::GPR64RegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(TargetOpcode::COPY), SrcReg).addReg(FramePtr);
|
2014-07-26 01:47:14 +08:00
|
|
|
// Recursively load frame address
|
|
|
|
// ldr x0, [fp]
|
|
|
|
// ldr x0, [x0]
|
|
|
|
// ldr x0, [x0]
|
|
|
|
// ...
|
|
|
|
unsigned DestReg;
|
|
|
|
unsigned Depth = cast<ConstantInt>(II->getOperand(0))->getZExtValue();
|
|
|
|
while (Depth--) {
|
2014-09-04 04:56:59 +08:00
|
|
|
DestReg = fastEmitInst_ri(AArch64::LDRXui, &AArch64::GPR64RegClass,
|
2014-08-22 04:57:57 +08:00
|
|
|
SrcReg, /*IsKill=*/true, 0);
|
|
|
|
assert(DestReg && "Unexpected LDR instruction emission failure.");
|
2014-07-26 01:47:14 +08:00
|
|
|
SrcReg = DestReg;
|
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(II, SrcReg);
|
2014-07-26 01:47:14 +08:00
|
|
|
return true;
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
case Intrinsic::memcpy:
|
|
|
|
case Intrinsic::memmove: {
|
2014-07-23 07:14:58 +08:00
|
|
|
const auto *MTI = cast<MemTransferInst>(II);
|
2014-03-29 18:18:08 +08:00
|
|
|
// Don't handle volatile.
|
2014-07-23 07:14:58 +08:00
|
|
|
if (MTI->isVolatile())
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-08-28 07:09:40 +08:00
|
|
|
// Disable inlining for memmove before calls to ComputeAddress. Otherwise,
|
2014-03-29 18:18:08 +08:00
|
|
|
// we would emit dead code because we don't currently handle memmoves.
|
2014-07-23 07:14:58 +08:00
|
|
|
bool IsMemCpy = (II->getIntrinsicID() == Intrinsic::memcpy);
|
|
|
|
if (isa<ConstantInt>(MTI->getLength()) && IsMemCpy) {
|
2014-03-29 18:18:08 +08:00
|
|
|
// Small memcpy's are common enough that we want to do them without a call
|
|
|
|
// if possible.
|
2014-07-23 07:14:58 +08:00
|
|
|
uint64_t Len = cast<ConstantInt>(MTI->getLength())->getZExtValue();
|
|
|
|
unsigned Alignment = MTI->getAlignment();
|
2014-09-16 07:20:17 +08:00
|
|
|
if (isMemCpySmall(Len, Alignment)) {
|
2014-03-29 18:18:08 +08:00
|
|
|
Address Dest, Src;
|
2014-09-16 07:20:17 +08:00
|
|
|
if (!computeAddress(MTI->getRawDest(), Dest) ||
|
|
|
|
!computeAddress(MTI->getRawSource(), Src))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
2014-09-16 07:20:17 +08:00
|
|
|
if (tryEmitSmallMemCpy(Dest, Src, Len, Alignment))
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
if (!MTI->getLength()->getType()->isIntegerTy(64))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
if (MTI->getSourceAddressSpace() > 255 || MTI->getDestAddressSpace() > 255)
|
2014-03-29 18:18:08 +08:00
|
|
|
// Fast instruction selection doesn't support the special
|
|
|
|
// address spaces.
|
|
|
|
return false;
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
const char *IntrMemName = isa<MemCpyInst>(II) ? "memcpy" : "memmove";
|
2014-09-04 04:56:52 +08:00
|
|
|
return lowerCallTo(II, IntrMemName, II->getNumArgOperands() - 2);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
case Intrinsic::memset: {
|
2014-07-23 07:14:58 +08:00
|
|
|
const MemSetInst *MSI = cast<MemSetInst>(II);
|
2014-03-29 18:18:08 +08:00
|
|
|
// Don't handle volatile.
|
2014-07-23 07:14:58 +08:00
|
|
|
if (MSI->isVolatile())
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
if (!MSI->getLength()->getType()->isIntegerTy(64))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-07-23 07:14:58 +08:00
|
|
|
if (MSI->getDestAddressSpace() > 255)
|
2014-03-29 18:18:08 +08:00
|
|
|
// Fast instruction selection doesn't support the special
|
|
|
|
// address spaces.
|
|
|
|
return false;
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
return lowerCallTo(II, "memset", II->getNumArgOperands() - 2);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
2014-09-16 06:33:06 +08:00
|
|
|
case Intrinsic::sin:
|
|
|
|
case Intrinsic::cos:
|
|
|
|
case Intrinsic::pow: {
|
|
|
|
MVT RetVT;
|
|
|
|
if (!isTypeLegal(II->getType(), RetVT))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (RetVT != MVT::f32 && RetVT != MVT::f64)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
static const RTLIB::Libcall LibCallTable[3][2] = {
|
|
|
|
{ RTLIB::SIN_F32, RTLIB::SIN_F64 },
|
|
|
|
{ RTLIB::COS_F32, RTLIB::COS_F64 },
|
|
|
|
{ RTLIB::POW_F32, RTLIB::POW_F64 }
|
|
|
|
};
|
|
|
|
RTLIB::Libcall LC;
|
|
|
|
bool Is64Bit = RetVT == MVT::f64;
|
|
|
|
switch (II->getIntrinsicID()) {
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unexpected intrinsic.");
|
|
|
|
case Intrinsic::sin:
|
|
|
|
LC = LibCallTable[0][Is64Bit];
|
|
|
|
break;
|
|
|
|
case Intrinsic::cos:
|
|
|
|
LC = LibCallTable[1][Is64Bit];
|
|
|
|
break;
|
|
|
|
case Intrinsic::pow:
|
|
|
|
LC = LibCallTable[2][Is64Bit];
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
ArgListTy Args;
|
|
|
|
Args.reserve(II->getNumArgOperands());
|
|
|
|
|
|
|
|
// Populate the argument list.
|
|
|
|
for (auto &Arg : II->arg_operands()) {
|
|
|
|
ArgListEntry Entry;
|
|
|
|
Entry.Val = Arg;
|
|
|
|
Entry.Ty = Arg->getType();
|
|
|
|
Args.push_back(Entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
CallLoweringInfo CLI;
|
|
|
|
CLI.setCallee(TLI.getLibcallCallingConv(LC), II->getType(),
|
|
|
|
TLI.getLibcallName(LC), std::move(Args));
|
|
|
|
if (!lowerCallTo(CLI))
|
|
|
|
return false;
|
|
|
|
updateValueMap(II, CLI.ResultReg);
|
|
|
|
return true;
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
case Intrinsic::trap: {
|
2014-05-24 20:50:23 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc, TII.get(AArch64::BRK))
|
2014-03-29 18:18:08 +08:00
|
|
|
.addImm(1);
|
|
|
|
return true;
|
|
|
|
}
|
2014-07-31 14:25:33 +08:00
|
|
|
case Intrinsic::sqrt: {
|
|
|
|
Type *RetTy = II->getCalledFunction()->getReturnType();
|
|
|
|
|
|
|
|
MVT VT;
|
|
|
|
if (!isTypeLegal(RetTy, VT))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned Op0Reg = getRegForValue(II->getOperand(0));
|
|
|
|
if (!Op0Reg)
|
|
|
|
return false;
|
|
|
|
bool Op0IsKill = hasTrivialKill(II->getOperand(0));
|
|
|
|
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned ResultReg = fastEmit_r(VT, VT, ISD::FSQRT, Op0Reg, Op0IsKill);
|
2014-07-31 14:25:33 +08:00
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(II, ResultReg);
|
2014-07-31 14:25:33 +08:00
|
|
|
return true;
|
|
|
|
}
|
2014-07-31 06:04:31 +08:00
|
|
|
case Intrinsic::sadd_with_overflow:
|
|
|
|
case Intrinsic::uadd_with_overflow:
|
|
|
|
case Intrinsic::ssub_with_overflow:
|
|
|
|
case Intrinsic::usub_with_overflow:
|
|
|
|
case Intrinsic::smul_with_overflow:
|
|
|
|
case Intrinsic::umul_with_overflow: {
|
|
|
|
// This implements the basic lowering of the xalu with overflow intrinsics.
|
|
|
|
const Function *Callee = II->getCalledFunction();
|
|
|
|
auto *Ty = cast<StructType>(Callee->getReturnType());
|
|
|
|
Type *RetTy = Ty->getTypeAtIndex(0U);
|
|
|
|
|
|
|
|
MVT VT;
|
|
|
|
if (!isTypeLegal(RetTy, VT))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (VT != MVT::i32 && VT != MVT::i64)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const Value *LHS = II->getArgOperand(0);
|
|
|
|
const Value *RHS = II->getArgOperand(1);
|
|
|
|
// Canonicalize immediate to the RHS.
|
|
|
|
if (isa<ConstantInt>(LHS) && !isa<ConstantInt>(RHS) &&
|
|
|
|
isCommutativeIntrinsic(II))
|
|
|
|
std::swap(LHS, RHS);
|
|
|
|
|
2014-09-18 15:04:54 +08:00
|
|
|
// Simplify multiplies.
|
|
|
|
unsigned IID = II->getIntrinsicID();
|
|
|
|
switch (IID) {
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
case Intrinsic::smul_with_overflow:
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(RHS))
|
|
|
|
if (C->getValue() == 2) {
|
|
|
|
IID = Intrinsic::sadd_with_overflow;
|
|
|
|
RHS = LHS;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case Intrinsic::umul_with_overflow:
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(RHS))
|
|
|
|
if (C->getValue() == 2) {
|
|
|
|
IID = Intrinsic::uadd_with_overflow;
|
|
|
|
RHS = LHS;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2014-08-20 06:29:55 +08:00
|
|
|
unsigned ResultReg1 = 0, ResultReg2 = 0, MulReg = 0;
|
2014-07-31 06:04:31 +08:00
|
|
|
AArch64CC::CondCode CC = AArch64CC::Invalid;
|
2014-09-18 15:04:54 +08:00
|
|
|
switch (IID) {
|
2014-07-31 06:04:31 +08:00
|
|
|
default: llvm_unreachable("Unexpected intrinsic!");
|
|
|
|
case Intrinsic::sadd_with_overflow:
|
2014-09-03 09:38:36 +08:00
|
|
|
ResultReg1 = emitAdd(VT, LHS, RHS, /*SetFlags=*/true);
|
|
|
|
CC = AArch64CC::VS;
|
|
|
|
break;
|
2014-07-31 06:04:31 +08:00
|
|
|
case Intrinsic::uadd_with_overflow:
|
2014-09-03 09:38:36 +08:00
|
|
|
ResultReg1 = emitAdd(VT, LHS, RHS, /*SetFlags=*/true);
|
|
|
|
CC = AArch64CC::HS;
|
|
|
|
break;
|
2014-07-31 06:04:31 +08:00
|
|
|
case Intrinsic::ssub_with_overflow:
|
2014-09-03 09:38:36 +08:00
|
|
|
ResultReg1 = emitSub(VT, LHS, RHS, /*SetFlags=*/true);
|
|
|
|
CC = AArch64CC::VS;
|
|
|
|
break;
|
2014-07-31 06:04:31 +08:00
|
|
|
case Intrinsic::usub_with_overflow:
|
2014-09-03 09:38:36 +08:00
|
|
|
ResultReg1 = emitSub(VT, LHS, RHS, /*SetFlags=*/true);
|
|
|
|
CC = AArch64CC::LO;
|
|
|
|
break;
|
2014-07-31 06:04:31 +08:00
|
|
|
case Intrinsic::smul_with_overflow: {
|
|
|
|
CC = AArch64CC::NE;
|
2014-08-20 06:29:55 +08:00
|
|
|
unsigned LHSReg = getRegForValue(LHS);
|
|
|
|
if (!LHSReg)
|
|
|
|
return false;
|
|
|
|
bool LHSIsKill = hasTrivialKill(LHS);
|
|
|
|
|
|
|
|
unsigned RHSReg = getRegForValue(RHS);
|
2014-08-01 09:25:55 +08:00
|
|
|
if (!RHSReg)
|
|
|
|
return false;
|
2014-08-20 06:29:55 +08:00
|
|
|
bool RHSIsKill = hasTrivialKill(RHS);
|
2014-08-01 09:25:55 +08:00
|
|
|
|
2014-07-31 06:04:31 +08:00
|
|
|
if (VT == MVT::i32) {
|
2014-09-16 07:20:17 +08:00
|
|
|
MulReg = emitSMULL_rr(MVT::i64, LHSReg, LHSIsKill, RHSReg, RHSIsKill);
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned ShiftReg = emitLSR_ri(MVT::i64, MVT::i64, MulReg,
|
|
|
|
/*IsKill=*/false, 32);
|
2014-09-04 04:56:59 +08:00
|
|
|
MulReg = fastEmitInst_extractsubreg(VT, MulReg, /*IsKill=*/true,
|
2014-07-31 06:04:31 +08:00
|
|
|
AArch64::sub_32);
|
2014-09-04 04:56:59 +08:00
|
|
|
ShiftReg = fastEmitInst_extractsubreg(VT, ShiftReg, /*IsKill=*/true,
|
2014-07-31 06:04:31 +08:00
|
|
|
AArch64::sub_32);
|
2014-08-20 06:29:55 +08:00
|
|
|
emitSubs_rs(VT, ShiftReg, /*IsKill=*/true, MulReg, /*IsKill=*/false,
|
|
|
|
AArch64_AM::ASR, 31, /*WantResult=*/false);
|
2014-07-31 06:04:31 +08:00
|
|
|
} else {
|
|
|
|
assert(VT == MVT::i64 && "Unexpected value type.");
|
2014-09-16 07:20:17 +08:00
|
|
|
MulReg = emitMul_rr(VT, LHSReg, LHSIsKill, RHSReg, RHSIsKill);
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned SMULHReg = fastEmit_rr(VT, VT, ISD::MULHS, LHSReg, LHSIsKill,
|
2014-07-31 06:04:31 +08:00
|
|
|
RHSReg, RHSIsKill);
|
2014-08-20 06:29:55 +08:00
|
|
|
emitSubs_rs(VT, SMULHReg, /*IsKill=*/true, MulReg, /*IsKill=*/false,
|
|
|
|
AArch64_AM::ASR, 63, /*WantResult=*/false);
|
2014-07-31 06:04:31 +08:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case Intrinsic::umul_with_overflow: {
|
|
|
|
CC = AArch64CC::NE;
|
2014-08-20 06:29:55 +08:00
|
|
|
unsigned LHSReg = getRegForValue(LHS);
|
|
|
|
if (!LHSReg)
|
|
|
|
return false;
|
|
|
|
bool LHSIsKill = hasTrivialKill(LHS);
|
|
|
|
|
|
|
|
unsigned RHSReg = getRegForValue(RHS);
|
2014-08-01 09:25:55 +08:00
|
|
|
if (!RHSReg)
|
|
|
|
return false;
|
2014-08-20 06:29:55 +08:00
|
|
|
bool RHSIsKill = hasTrivialKill(RHS);
|
2014-08-01 09:25:55 +08:00
|
|
|
|
2014-07-31 06:04:31 +08:00
|
|
|
if (VT == MVT::i32) {
|
2014-09-16 07:20:17 +08:00
|
|
|
MulReg = emitUMULL_rr(MVT::i64, LHSReg, LHSIsKill, RHSReg, RHSIsKill);
|
2014-08-20 06:29:55 +08:00
|
|
|
emitSubs_rs(MVT::i64, AArch64::XZR, /*IsKill=*/true, MulReg,
|
|
|
|
/*IsKill=*/false, AArch64_AM::LSR, 32,
|
|
|
|
/*WantResult=*/false);
|
2014-09-04 04:56:59 +08:00
|
|
|
MulReg = fastEmitInst_extractsubreg(VT, MulReg, /*IsKill=*/true,
|
2014-07-31 06:04:31 +08:00
|
|
|
AArch64::sub_32);
|
|
|
|
} else {
|
|
|
|
assert(VT == MVT::i64 && "Unexpected value type.");
|
2014-09-16 07:20:17 +08:00
|
|
|
MulReg = emitMul_rr(VT, LHSReg, LHSIsKill, RHSReg, RHSIsKill);
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned UMULHReg = fastEmit_rr(VT, VT, ISD::MULHU, LHSReg, LHSIsKill,
|
2014-07-31 06:04:31 +08:00
|
|
|
RHSReg, RHSIsKill);
|
2014-08-20 06:29:55 +08:00
|
|
|
emitSubs_rr(VT, AArch64::XZR, /*IsKill=*/true, UMULHReg,
|
|
|
|
/*IsKill=*/false, /*WantResult=*/false);
|
2014-07-31 06:04:31 +08:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-08-20 06:29:55 +08:00
|
|
|
if (MulReg) {
|
|
|
|
ResultReg1 = createResultReg(TLI.getRegClassFor(VT));
|
2014-07-31 06:04:31 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
2014-08-20 06:29:55 +08:00
|
|
|
TII.get(TargetOpcode::COPY), ResultReg1).addReg(MulReg);
|
|
|
|
}
|
2014-07-31 06:04:31 +08:00
|
|
|
|
2014-09-04 04:56:59 +08:00
|
|
|
ResultReg2 = fastEmitInst_rri(AArch64::CSINCWr, &AArch64::GPR32RegClass,
|
2014-08-22 04:57:57 +08:00
|
|
|
AArch64::WZR, /*IsKill=*/true, AArch64::WZR,
|
|
|
|
/*IsKill=*/true, getInvertedCondCode(CC));
|
2014-08-20 06:29:55 +08:00
|
|
|
assert((ResultReg1 + 1) == ResultReg2 &&
|
|
|
|
"Nonconsecutive result registers.");
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(II, ResultReg1, 2);
|
2014-07-31 06:04:31 +08:00
|
|
|
return true;
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectRet(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
const ReturnInst *Ret = cast<ReturnInst>(I);
|
|
|
|
const Function &F = *I->getParent()->getParent();
|
|
|
|
|
|
|
|
if (!FuncInfo.CanLowerReturn)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (F.isVarArg())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Build a list of return value registers.
|
|
|
|
SmallVector<unsigned, 4> RetRegs;
|
|
|
|
|
|
|
|
if (Ret->getNumOperands() > 0) {
|
|
|
|
CallingConv::ID CC = F.getCallingConv();
|
|
|
|
SmallVector<ISD::OutputArg, 4> Outs;
|
|
|
|
GetReturnInfo(F.getReturnType(), F.getAttributes(), Outs, TLI);
|
|
|
|
|
|
|
|
// Analyze operands of the call, assigning locations to each operand.
|
|
|
|
SmallVector<CCValAssign, 16> ValLocs;
|
2014-08-07 02:45:26 +08:00
|
|
|
CCState CCInfo(CC, F.isVarArg(), *FuncInfo.MF, ValLocs, I->getContext());
|
2014-05-24 20:50:23 +08:00
|
|
|
CCAssignFn *RetCC = CC == CallingConv::WebKit_JS ? RetCC_AArch64_WebKit_JS
|
|
|
|
: RetCC_AArch64_AAPCS;
|
2014-03-29 18:18:08 +08:00
|
|
|
CCInfo.AnalyzeReturn(Outs, RetCC);
|
|
|
|
|
|
|
|
// Only handle a single return value for now.
|
|
|
|
if (ValLocs.size() != 1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
CCValAssign &VA = ValLocs[0];
|
|
|
|
const Value *RV = Ret->getOperand(0);
|
|
|
|
|
|
|
|
// Don't bother handling odd stuff for now.
|
2014-09-16 07:40:10 +08:00
|
|
|
if ((VA.getLocInfo() != CCValAssign::Full) &&
|
|
|
|
(VA.getLocInfo() != CCValAssign::BCvt))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
2014-09-16 07:40:10 +08:00
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
// Only handle register returns for now.
|
|
|
|
if (!VA.isRegLoc())
|
|
|
|
return false;
|
2014-09-16 07:40:10 +08:00
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned Reg = getRegForValue(RV);
|
|
|
|
if (Reg == 0)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned SrcReg = Reg + VA.getValNo();
|
|
|
|
unsigned DestReg = VA.getLocReg();
|
|
|
|
// Avoid a cross-class copy. This is very unlikely.
|
|
|
|
if (!MRI.getRegClass(SrcReg)->contains(DestReg))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
EVT RVEVT = TLI.getValueType(RV->getType());
|
|
|
|
if (!RVEVT.isSimple())
|
|
|
|
return false;
|
2014-05-07 20:33:55 +08:00
|
|
|
|
|
|
|
// Vectors (of > 1 lane) in big endian need tricky handling.
|
2014-09-16 07:40:10 +08:00
|
|
|
if (RVEVT.isVector() && RVEVT.getVectorNumElements() > 1 &&
|
|
|
|
!Subtarget->isLittleEndian())
|
2014-05-07 20:33:55 +08:00
|
|
|
return false;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
MVT RVVT = RVEVT.getSimpleVT();
|
2014-04-30 23:29:57 +08:00
|
|
|
if (RVVT == MVT::f128)
|
|
|
|
return false;
|
2014-09-16 07:40:10 +08:00
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
MVT DestVT = VA.getValVT();
|
|
|
|
// Special handling for extended integers.
|
|
|
|
if (RVVT != DestVT) {
|
|
|
|
if (RVVT != MVT::i1 && RVVT != MVT::i8 && RVVT != MVT::i16)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (!Outs[0].Flags.isZExt() && !Outs[0].Flags.isSExt())
|
|
|
|
return false;
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool IsZExt = Outs[0].Flags.isZExt();
|
|
|
|
SrcReg = emitIntExt(RVVT, SrcReg, DestVT, IsZExt);
|
2014-03-29 18:18:08 +08:00
|
|
|
if (SrcReg == 0)
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make the copy.
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(TargetOpcode::COPY), DestReg).addReg(SrcReg);
|
|
|
|
|
|
|
|
// Add register to return instruction.
|
|
|
|
RetRegs.push_back(VA.getLocReg());
|
|
|
|
}
|
|
|
|
|
|
|
|
MachineInstrBuilder MIB = BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
2014-05-24 20:50:23 +08:00
|
|
|
TII.get(AArch64::RET_ReallyLR));
|
2014-03-29 18:18:08 +08:00
|
|
|
for (unsigned i = 0, e = RetRegs.size(); i != e; ++i)
|
|
|
|
MIB.addReg(RetRegs[i], RegState::Implicit);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectTrunc(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
Type *DestTy = I->getType();
|
|
|
|
Value *Op = I->getOperand(0);
|
|
|
|
Type *SrcTy = Op->getType();
|
|
|
|
|
|
|
|
EVT SrcEVT = TLI.getValueType(SrcTy, true);
|
|
|
|
EVT DestEVT = TLI.getValueType(DestTy, true);
|
|
|
|
if (!SrcEVT.isSimple())
|
|
|
|
return false;
|
|
|
|
if (!DestEVT.isSimple())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MVT SrcVT = SrcEVT.getSimpleVT();
|
|
|
|
MVT DestVT = DestEVT.getSimpleVT();
|
|
|
|
|
|
|
|
if (SrcVT != MVT::i64 && SrcVT != MVT::i32 && SrcVT != MVT::i16 &&
|
|
|
|
SrcVT != MVT::i8)
|
|
|
|
return false;
|
|
|
|
if (DestVT != MVT::i32 && DestVT != MVT::i16 && DestVT != MVT::i8 &&
|
|
|
|
DestVT != MVT::i1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned SrcReg = getRegForValue(Op);
|
|
|
|
if (!SrcReg)
|
|
|
|
return false;
|
2014-08-22 02:02:25 +08:00
|
|
|
bool SrcIsKill = hasTrivialKill(Op);
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
// If we're truncating from i64 to a smaller non-legal type then generate an
|
2014-08-30 01:58:16 +08:00
|
|
|
// AND. Otherwise, we know the high bits are undefined and a truncate only
|
|
|
|
// generate a COPY. We cannot mark the source register also as result
|
|
|
|
// register, because this can incorrectly transfer the kill flag onto the
|
|
|
|
// source register.
|
|
|
|
unsigned ResultReg;
|
2014-03-29 18:18:08 +08:00
|
|
|
if (SrcVT == MVT::i64) {
|
|
|
|
uint64_t Mask = 0;
|
|
|
|
switch (DestVT.SimpleTy) {
|
|
|
|
default:
|
|
|
|
// Trunc i64 to i32 is handled by the target-independent fast-isel.
|
|
|
|
return false;
|
|
|
|
case MVT::i1:
|
|
|
|
Mask = 0x1;
|
|
|
|
break;
|
|
|
|
case MVT::i8:
|
|
|
|
Mask = 0xff;
|
|
|
|
break;
|
|
|
|
case MVT::i16:
|
|
|
|
Mask = 0xffff;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
// Issue an extract_subreg to get the lower 32-bits.
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned Reg32 = fastEmitInst_extractsubreg(MVT::i32, SrcReg, SrcIsKill,
|
2014-05-24 20:50:23 +08:00
|
|
|
AArch64::sub_32);
|
2014-03-29 18:18:08 +08:00
|
|
|
// Create the AND instruction which performs the actual truncation.
|
2014-09-04 09:29:18 +08:00
|
|
|
ResultReg = emitAnd_ri(MVT::i32, Reg32, /*IsKill=*/true, Mask);
|
2014-08-30 01:58:16 +08:00
|
|
|
assert(ResultReg && "Unexpected AND instruction emission failure.");
|
|
|
|
} else {
|
|
|
|
ResultReg = createResultReg(&AArch64::GPR32RegClass);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(TargetOpcode::COPY), ResultReg)
|
|
|
|
.addReg(SrcReg, getKillRegState(SrcIsKill));
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned AArch64FastISel::emiti1Ext(unsigned SrcReg, MVT DestVT, bool IsZExt) {
|
2014-03-29 18:18:08 +08:00
|
|
|
assert((DestVT == MVT::i8 || DestVT == MVT::i16 || DestVT == MVT::i32 ||
|
|
|
|
DestVT == MVT::i64) &&
|
|
|
|
"Unexpected value type.");
|
|
|
|
// Handle i8 and i16 as i32.
|
|
|
|
if (DestVT == MVT::i8 || DestVT == MVT::i16)
|
|
|
|
DestVT = MVT::i32;
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
if (IsZExt) {
|
2014-09-04 09:29:18 +08:00
|
|
|
unsigned ResultReg = emitAnd_ri(MVT::i32, SrcReg, /*TODO:IsKill=*/false, 1);
|
2014-08-22 02:02:25 +08:00
|
|
|
assert(ResultReg && "Unexpected AND instruction emission failure.");
|
2014-03-29 18:18:08 +08:00
|
|
|
if (DestVT == MVT::i64) {
|
|
|
|
// We're ZExt i1 to i64. The ANDWri Wd, Ws, #1 implicitly clears the
|
|
|
|
// upper 32 bits. Emit a SUBREG_TO_REG to extend from Wd to Xd.
|
2014-05-24 20:50:23 +08:00
|
|
|
unsigned Reg64 = MRI.createVirtualRegister(&AArch64::GPR64RegClass);
|
2014-03-29 18:18:08 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
2014-05-24 20:50:23 +08:00
|
|
|
TII.get(AArch64::SUBREG_TO_REG), Reg64)
|
2014-03-29 18:18:08 +08:00
|
|
|
.addImm(0)
|
|
|
|
.addReg(ResultReg)
|
2014-05-24 20:50:23 +08:00
|
|
|
.addImm(AArch64::sub_32);
|
2014-03-29 18:18:08 +08:00
|
|
|
ResultReg = Reg64;
|
|
|
|
}
|
|
|
|
return ResultReg;
|
|
|
|
} else {
|
|
|
|
if (DestVT == MVT::i64) {
|
|
|
|
// FIXME: We're SExt i1 to i64.
|
|
|
|
return 0;
|
|
|
|
}
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_rii(AArch64::SBFMWri, &AArch64::GPR32RegClass, SrcReg,
|
2014-08-22 04:57:57 +08:00
|
|
|
/*TODO:IsKill=*/false, 0, 0);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned AArch64FastISel::emitMul_rr(MVT RetVT, unsigned Op0, bool Op0IsKill,
|
2014-07-31 06:04:25 +08:00
|
|
|
unsigned Op1, bool Op1IsKill) {
|
|
|
|
unsigned Opc, ZReg;
|
|
|
|
switch (RetVT.SimpleTy) {
|
|
|
|
default: return 0;
|
|
|
|
case MVT::i8:
|
|
|
|
case MVT::i16:
|
|
|
|
case MVT::i32:
|
|
|
|
RetVT = MVT::i32;
|
|
|
|
Opc = AArch64::MADDWrrr; ZReg = AArch64::WZR; break;
|
|
|
|
case MVT::i64:
|
|
|
|
Opc = AArch64::MADDXrrr; ZReg = AArch64::XZR; break;
|
|
|
|
}
|
|
|
|
|
2014-08-22 04:57:57 +08:00
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
(RetVT == MVT::i64) ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_rrr(Opc, RC, Op0, Op0IsKill, Op1, Op1IsKill,
|
2014-08-22 04:57:57 +08:00
|
|
|
/*IsKill=*/ZReg, true);
|
2014-07-31 06:04:25 +08:00
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned AArch64FastISel::emitSMULL_rr(MVT RetVT, unsigned Op0, bool Op0IsKill,
|
2014-07-31 06:04:25 +08:00
|
|
|
unsigned Op1, bool Op1IsKill) {
|
|
|
|
if (RetVT != MVT::i64)
|
|
|
|
return 0;
|
|
|
|
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_rrr(AArch64::SMADDLrrr, &AArch64::GPR64RegClass,
|
2014-08-22 04:57:57 +08:00
|
|
|
Op0, Op0IsKill, Op1, Op1IsKill,
|
|
|
|
AArch64::XZR, /*IsKill=*/true);
|
2014-07-31 06:04:25 +08:00
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned AArch64FastISel::emitUMULL_rr(MVT RetVT, unsigned Op0, bool Op0IsKill,
|
2014-07-31 06:04:25 +08:00
|
|
|
unsigned Op1, bool Op1IsKill) {
|
|
|
|
if (RetVT != MVT::i64)
|
|
|
|
return 0;
|
|
|
|
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_rrr(AArch64::UMADDLrrr, &AArch64::GPR64RegClass,
|
2014-08-22 04:57:57 +08:00
|
|
|
Op0, Op0IsKill, Op1, Op1IsKill,
|
|
|
|
AArch64::XZR, /*IsKill=*/true);
|
2014-07-31 06:04:25 +08:00
|
|
|
}
|
|
|
|
|
2014-08-22 07:06:07 +08:00
|
|
|
unsigned AArch64FastISel::emitLSL_rr(MVT RetVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
unsigned Op1Reg, bool Op1IsKill) {
|
|
|
|
unsigned Opc = 0;
|
|
|
|
bool NeedTrunc = false;
|
|
|
|
uint64_t Mask = 0;
|
|
|
|
switch (RetVT.SimpleTy) {
|
|
|
|
default: return 0;
|
|
|
|
case MVT::i8: Opc = AArch64::LSLVWr; NeedTrunc = true; Mask = 0xff; break;
|
|
|
|
case MVT::i16: Opc = AArch64::LSLVWr; NeedTrunc = true; Mask = 0xffff; break;
|
|
|
|
case MVT::i32: Opc = AArch64::LSLVWr; break;
|
|
|
|
case MVT::i64: Opc = AArch64::LSLVXr; break;
|
|
|
|
}
|
|
|
|
|
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
(RetVT == MVT::i64) ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
|
|
|
if (NeedTrunc) {
|
2014-09-04 09:29:18 +08:00
|
|
|
Op1Reg = emitAnd_ri(MVT::i32, Op1Reg, Op1IsKill, Mask);
|
2014-08-22 07:06:07 +08:00
|
|
|
Op1IsKill = true;
|
|
|
|
}
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned ResultReg = fastEmitInst_rr(Opc, RC, Op0Reg, Op0IsKill, Op1Reg,
|
2014-08-22 07:06:07 +08:00
|
|
|
Op1IsKill);
|
|
|
|
if (NeedTrunc)
|
2014-09-04 09:29:18 +08:00
|
|
|
ResultReg = emitAnd_ri(MVT::i32, ResultReg, /*IsKill=*/true, Mask);
|
2014-08-22 07:06:07 +08:00
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned AArch64FastISel::emitLSL_ri(MVT RetVT, MVT SrcVT, unsigned Op0,
|
|
|
|
bool Op0IsKill, uint64_t Shift,
|
|
|
|
bool IsZext) {
|
|
|
|
assert(RetVT.SimpleTy >= SrcVT.SimpleTy &&
|
|
|
|
"Unexpected source/return type pair.");
|
2014-09-23 05:08:53 +08:00
|
|
|
assert((SrcVT == MVT::i1 || SrcVT == MVT::i8 || SrcVT == MVT::i16 ||
|
|
|
|
SrcVT == MVT::i32 || SrcVT == MVT::i64) &&
|
|
|
|
"Unexpected source value type.");
|
2014-08-27 08:58:26 +08:00
|
|
|
assert((RetVT == MVT::i8 || RetVT == MVT::i16 || RetVT == MVT::i32 ||
|
|
|
|
RetVT == MVT::i64) && "Unexpected return value type.");
|
|
|
|
|
|
|
|
bool Is64Bit = (RetVT == MVT::i64);
|
|
|
|
unsigned RegSize = Is64Bit ? 64 : 32;
|
|
|
|
unsigned DstBits = RetVT.getSizeInBits();
|
|
|
|
unsigned SrcBits = SrcVT.getSizeInBits();
|
2014-07-31 06:04:22 +08:00
|
|
|
|
2014-08-27 08:58:26 +08:00
|
|
|
// Don't deal with undefined shifts.
|
|
|
|
if (Shift >= DstBits)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
// For immediate shifts we can fold the zero-/sign-extension into the shift.
|
|
|
|
// {S|U}BFM Wd, Wn, #r, #s
|
|
|
|
// Wd<32+s-r,32-r> = Wn<s:0> when r > s
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = shl i16 %1, 4
|
|
|
|
// Wd<32+7-28,32-28> = Wn<7:0> <- clamp s to 7
|
|
|
|
// 0b1111_1111_1111_1111__1111_1010_1010_0000 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0101_0101_0000 sext | zext
|
|
|
|
// 0b0000_0000_0000_0000__0000_1010_1010_0000 zext
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = shl i16 %1, 8
|
|
|
|
// Wd<32+7-24,32-24> = Wn<7:0>
|
|
|
|
// 0b1111_1111_1111_1111__1010_1010_0000_0000 sext
|
|
|
|
// 0b0000_0000_0000_0000__0101_0101_0000_0000 sext | zext
|
|
|
|
// 0b0000_0000_0000_0000__1010_1010_0000_0000 zext
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = shl i16 %1, 12
|
|
|
|
// Wd<32+3-20,32-20> = Wn<3:0>
|
|
|
|
// 0b1111_1111_1111_1111__1010_0000_0000_0000 sext
|
|
|
|
// 0b0000_0000_0000_0000__0101_0000_0000_0000 sext | zext
|
|
|
|
// 0b0000_0000_0000_0000__1010_0000_0000_0000 zext
|
|
|
|
|
|
|
|
unsigned ImmR = RegSize - Shift;
|
|
|
|
// Limit the width to the length of the source type.
|
|
|
|
unsigned ImmS = std::min<unsigned>(SrcBits - 1, DstBits - 1 - Shift);
|
|
|
|
static const unsigned OpcTable[2][2] = {
|
|
|
|
{AArch64::SBFMWri, AArch64::SBFMXri},
|
|
|
|
{AArch64::UBFMWri, AArch64::UBFMXri}
|
|
|
|
};
|
|
|
|
unsigned Opc = OpcTable[IsZext][Is64Bit];
|
2014-08-22 04:57:57 +08:00
|
|
|
const TargetRegisterClass *RC =
|
2014-08-27 08:58:26 +08:00
|
|
|
Is64Bit ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
|
|
|
if (SrcVT.SimpleTy <= MVT::i32 && RetVT == MVT::i64) {
|
|
|
|
unsigned TmpReg = MRI.createVirtualRegister(RC);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(AArch64::SUBREG_TO_REG), TmpReg)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(Op0, getKillRegState(Op0IsKill))
|
|
|
|
.addImm(AArch64::sub_32);
|
|
|
|
Op0 = TmpReg;
|
|
|
|
Op0IsKill = true;
|
|
|
|
}
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_rii(Opc, RC, Op0, Op0IsKill, ImmR, ImmS);
|
2014-07-31 06:04:22 +08:00
|
|
|
}
|
|
|
|
|
2014-08-22 07:06:07 +08:00
|
|
|
unsigned AArch64FastISel::emitLSR_rr(MVT RetVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
unsigned Op1Reg, bool Op1IsKill) {
|
|
|
|
unsigned Opc = 0;
|
|
|
|
bool NeedTrunc = false;
|
|
|
|
uint64_t Mask = 0;
|
|
|
|
switch (RetVT.SimpleTy) {
|
|
|
|
default: return 0;
|
|
|
|
case MVT::i8: Opc = AArch64::LSRVWr; NeedTrunc = true; Mask = 0xff; break;
|
|
|
|
case MVT::i16: Opc = AArch64::LSRVWr; NeedTrunc = true; Mask = 0xffff; break;
|
|
|
|
case MVT::i32: Opc = AArch64::LSRVWr; break;
|
|
|
|
case MVT::i64: Opc = AArch64::LSRVXr; break;
|
|
|
|
}
|
|
|
|
|
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
(RetVT == MVT::i64) ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
|
|
|
if (NeedTrunc) {
|
2014-09-04 09:29:18 +08:00
|
|
|
Op0Reg = emitAnd_ri(MVT::i32, Op0Reg, Op0IsKill, Mask);
|
|
|
|
Op1Reg = emitAnd_ri(MVT::i32, Op1Reg, Op1IsKill, Mask);
|
2014-08-22 07:06:07 +08:00
|
|
|
Op0IsKill = Op1IsKill = true;
|
|
|
|
}
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned ResultReg = fastEmitInst_rr(Opc, RC, Op0Reg, Op0IsKill, Op1Reg,
|
2014-08-22 07:06:07 +08:00
|
|
|
Op1IsKill);
|
|
|
|
if (NeedTrunc)
|
2014-09-04 09:29:18 +08:00
|
|
|
ResultReg = emitAnd_ri(MVT::i32, ResultReg, /*IsKill=*/true, Mask);
|
2014-08-22 07:06:07 +08:00
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned AArch64FastISel::emitLSR_ri(MVT RetVT, MVT SrcVT, unsigned Op0,
|
|
|
|
bool Op0IsKill, uint64_t Shift,
|
|
|
|
bool IsZExt) {
|
|
|
|
assert(RetVT.SimpleTy >= SrcVT.SimpleTy &&
|
|
|
|
"Unexpected source/return type pair.");
|
|
|
|
assert((SrcVT == MVT::i8 || SrcVT == MVT::i16 || SrcVT == MVT::i32 ||
|
|
|
|
SrcVT == MVT::i64) && "Unexpected source value type.");
|
|
|
|
assert((RetVT == MVT::i8 || RetVT == MVT::i16 || RetVT == MVT::i32 ||
|
|
|
|
RetVT == MVT::i64) && "Unexpected return value type.");
|
|
|
|
|
|
|
|
bool Is64Bit = (RetVT == MVT::i64);
|
|
|
|
unsigned RegSize = Is64Bit ? 64 : 32;
|
|
|
|
unsigned DstBits = RetVT.getSizeInBits();
|
|
|
|
unsigned SrcBits = SrcVT.getSizeInBits();
|
|
|
|
|
|
|
|
// Don't deal with undefined shifts.
|
|
|
|
if (Shift >= DstBits)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
// For immediate shifts we can fold the zero-/sign-extension into the shift.
|
|
|
|
// {S|U}BFM Wd, Wn, #r, #s
|
|
|
|
// Wd<s-r:0> = Wn<s:r> when r <= s
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = lshr i16 %1, 4
|
|
|
|
// Wd<7-4:0> = Wn<7:4>
|
|
|
|
// 0b0000_0000_0000_0000__0000_1111_1111_1010 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0101 sext | zext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_1010 zext
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = lshr i16 %1, 8
|
|
|
|
// Wd<7-7,0> = Wn<7:7>
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_1111_1111 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0000 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0000 zext
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = lshr i16 %1, 12
|
|
|
|
// Wd<7-7,0> = Wn<7:7> <- clamp r to 7
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_1111 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0000 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0000 zext
|
|
|
|
|
|
|
|
if (Shift >= SrcBits && IsZExt)
|
2014-09-16 07:20:17 +08:00
|
|
|
return materializeInt(ConstantInt::get(*Context, APInt(RegSize, 0)), RetVT);
|
2014-08-27 08:58:26 +08:00
|
|
|
|
|
|
|
// It is not possible to fold a sign-extend into the LShr instruction. In this
|
|
|
|
// case emit a sign-extend.
|
|
|
|
if (!IsZExt) {
|
2014-09-16 07:20:17 +08:00
|
|
|
Op0 = emitIntExt(SrcVT, Op0, RetVT, IsZExt);
|
2014-08-27 08:58:26 +08:00
|
|
|
if (!Op0)
|
|
|
|
return 0;
|
|
|
|
Op0IsKill = true;
|
|
|
|
SrcVT = RetVT;
|
|
|
|
SrcBits = SrcVT.getSizeInBits();
|
|
|
|
IsZExt = true;
|
2014-07-31 06:04:22 +08:00
|
|
|
}
|
|
|
|
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned ImmR = std::min<unsigned>(SrcBits - 1, Shift);
|
|
|
|
unsigned ImmS = SrcBits - 1;
|
|
|
|
static const unsigned OpcTable[2][2] = {
|
|
|
|
{AArch64::SBFMWri, AArch64::SBFMXri},
|
|
|
|
{AArch64::UBFMWri, AArch64::UBFMXri}
|
|
|
|
};
|
|
|
|
unsigned Opc = OpcTable[IsZExt][Is64Bit];
|
2014-08-22 04:57:57 +08:00
|
|
|
const TargetRegisterClass *RC =
|
2014-08-27 08:58:26 +08:00
|
|
|
Is64Bit ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
|
|
|
if (SrcVT.SimpleTy <= MVT::i32 && RetVT == MVT::i64) {
|
|
|
|
unsigned TmpReg = MRI.createVirtualRegister(RC);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(AArch64::SUBREG_TO_REG), TmpReg)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(Op0, getKillRegState(Op0IsKill))
|
|
|
|
.addImm(AArch64::sub_32);
|
|
|
|
Op0 = TmpReg;
|
|
|
|
Op0IsKill = true;
|
|
|
|
}
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_rii(Opc, RC, Op0, Op0IsKill, ImmR, ImmS);
|
2014-07-31 06:04:22 +08:00
|
|
|
}
|
|
|
|
|
2014-08-22 07:06:07 +08:00
|
|
|
unsigned AArch64FastISel::emitASR_rr(MVT RetVT, unsigned Op0Reg, bool Op0IsKill,
|
|
|
|
unsigned Op1Reg, bool Op1IsKill) {
|
|
|
|
unsigned Opc = 0;
|
|
|
|
bool NeedTrunc = false;
|
|
|
|
uint64_t Mask = 0;
|
|
|
|
switch (RetVT.SimpleTy) {
|
|
|
|
default: return 0;
|
|
|
|
case MVT::i8: Opc = AArch64::ASRVWr; NeedTrunc = true; Mask = 0xff; break;
|
|
|
|
case MVT::i16: Opc = AArch64::ASRVWr; NeedTrunc = true; Mask = 0xffff; break;
|
|
|
|
case MVT::i32: Opc = AArch64::ASRVWr; break;
|
|
|
|
case MVT::i64: Opc = AArch64::ASRVXr; break;
|
|
|
|
}
|
|
|
|
|
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
(RetVT == MVT::i64) ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
|
|
|
if (NeedTrunc) {
|
2014-09-16 07:20:17 +08:00
|
|
|
Op0Reg = emitIntExt(RetVT, Op0Reg, MVT::i32, /*IsZExt=*/false);
|
2014-09-04 09:29:18 +08:00
|
|
|
Op1Reg = emitAnd_ri(MVT::i32, Op1Reg, Op1IsKill, Mask);
|
2014-08-22 07:06:07 +08:00
|
|
|
Op0IsKill = Op1IsKill = true;
|
|
|
|
}
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned ResultReg = fastEmitInst_rr(Opc, RC, Op0Reg, Op0IsKill, Op1Reg,
|
2014-08-22 07:06:07 +08:00
|
|
|
Op1IsKill);
|
|
|
|
if (NeedTrunc)
|
2014-09-04 09:29:18 +08:00
|
|
|
ResultReg = emitAnd_ri(MVT::i32, ResultReg, /*IsKill=*/true, Mask);
|
2014-08-22 07:06:07 +08:00
|
|
|
return ResultReg;
|
|
|
|
}
|
|
|
|
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned AArch64FastISel::emitASR_ri(MVT RetVT, MVT SrcVT, unsigned Op0,
|
|
|
|
bool Op0IsKill, uint64_t Shift,
|
|
|
|
bool IsZExt) {
|
|
|
|
assert(RetVT.SimpleTy >= SrcVT.SimpleTy &&
|
|
|
|
"Unexpected source/return type pair.");
|
|
|
|
assert((SrcVT == MVT::i8 || SrcVT == MVT::i16 || SrcVT == MVT::i32 ||
|
|
|
|
SrcVT == MVT::i64) && "Unexpected source value type.");
|
|
|
|
assert((RetVT == MVT::i8 || RetVT == MVT::i16 || RetVT == MVT::i32 ||
|
|
|
|
RetVT == MVT::i64) && "Unexpected return value type.");
|
|
|
|
|
|
|
|
bool Is64Bit = (RetVT == MVT::i64);
|
|
|
|
unsigned RegSize = Is64Bit ? 64 : 32;
|
|
|
|
unsigned DstBits = RetVT.getSizeInBits();
|
|
|
|
unsigned SrcBits = SrcVT.getSizeInBits();
|
2014-07-31 06:04:22 +08:00
|
|
|
|
2014-08-27 08:58:26 +08:00
|
|
|
// Don't deal with undefined shifts.
|
|
|
|
if (Shift >= DstBits)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
// For immediate shifts we can fold the zero-/sign-extension into the shift.
|
|
|
|
// {S|U}BFM Wd, Wn, #r, #s
|
|
|
|
// Wd<s-r:0> = Wn<s:r> when r <= s
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = ashr i16 %1, 4
|
|
|
|
// Wd<7-4:0> = Wn<7:4>
|
|
|
|
// 0b1111_1111_1111_1111__1111_1111_1111_1010 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0101 sext | zext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_1010 zext
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = ashr i16 %1, 8
|
|
|
|
// Wd<7-7,0> = Wn<7:7>
|
|
|
|
// 0b1111_1111_1111_1111__1111_1111_1111_1111 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0000 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0000 zext
|
|
|
|
|
|
|
|
// %1 = {s|z}ext i8 {0b1010_1010|0b0101_0101} to i16
|
|
|
|
// %2 = ashr i16 %1, 12
|
|
|
|
// Wd<7-7,0> = Wn<7:7> <- clamp r to 7
|
|
|
|
// 0b1111_1111_1111_1111__1111_1111_1111_1111 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0000 sext
|
|
|
|
// 0b0000_0000_0000_0000__0000_0000_0000_0000 zext
|
|
|
|
|
|
|
|
if (Shift >= SrcBits && IsZExt)
|
2014-09-16 07:20:17 +08:00
|
|
|
return materializeInt(ConstantInt::get(*Context, APInt(RegSize, 0)), RetVT);
|
2014-08-27 08:58:26 +08:00
|
|
|
|
|
|
|
unsigned ImmR = std::min<unsigned>(SrcBits - 1, Shift);
|
|
|
|
unsigned ImmS = SrcBits - 1;
|
|
|
|
static const unsigned OpcTable[2][2] = {
|
|
|
|
{AArch64::SBFMWri, AArch64::SBFMXri},
|
|
|
|
{AArch64::UBFMWri, AArch64::UBFMXri}
|
|
|
|
};
|
|
|
|
unsigned Opc = OpcTable[IsZExt][Is64Bit];
|
2014-08-22 04:57:57 +08:00
|
|
|
const TargetRegisterClass *RC =
|
2014-08-27 08:58:26 +08:00
|
|
|
Is64Bit ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
|
|
|
if (SrcVT.SimpleTy <= MVT::i32 && RetVT == MVT::i64) {
|
|
|
|
unsigned TmpReg = MRI.createVirtualRegister(RC);
|
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(AArch64::SUBREG_TO_REG), TmpReg)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(Op0, getKillRegState(Op0IsKill))
|
|
|
|
.addImm(AArch64::sub_32);
|
|
|
|
Op0 = TmpReg;
|
|
|
|
Op0IsKill = true;
|
|
|
|
}
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_rii(Opc, RC, Op0, Op0IsKill, ImmR, ImmS);
|
2014-07-31 06:04:22 +08:00
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned AArch64FastISel::emitIntExt(MVT SrcVT, unsigned SrcReg, MVT DestVT,
|
|
|
|
bool IsZExt) {
|
2014-03-29 18:18:08 +08:00
|
|
|
assert(DestVT != MVT::i1 && "ZeroExt/SignExt an i1?");
|
2014-07-08 05:37:51 +08:00
|
|
|
|
2014-07-10 01:54:32 +08:00
|
|
|
// FastISel does not have plumbing to deal with extensions where the SrcVT or
|
|
|
|
// DestVT are odd things, so test to make sure that they are both types we can
|
|
|
|
// handle (i1/i8/i16/i32 for SrcVT and i8/i16/i32/i64 for DestVT), otherwise
|
|
|
|
// bail out to SelectionDAG.
|
|
|
|
if (((DestVT != MVT::i8) && (DestVT != MVT::i16) &&
|
|
|
|
(DestVT != MVT::i32) && (DestVT != MVT::i64)) ||
|
|
|
|
((SrcVT != MVT::i1) && (SrcVT != MVT::i8) &&
|
|
|
|
(SrcVT != MVT::i16) && (SrcVT != MVT::i32)))
|
2014-07-08 05:37:51 +08:00
|
|
|
return 0;
|
|
|
|
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned Opc;
|
|
|
|
unsigned Imm = 0;
|
|
|
|
|
|
|
|
switch (SrcVT.SimpleTy) {
|
|
|
|
default:
|
|
|
|
return 0;
|
|
|
|
case MVT::i1:
|
2014-09-16 07:20:17 +08:00
|
|
|
return emiti1Ext(SrcReg, DestVT, IsZExt);
|
2014-03-29 18:18:08 +08:00
|
|
|
case MVT::i8:
|
|
|
|
if (DestVT == MVT::i64)
|
2014-09-16 07:20:17 +08:00
|
|
|
Opc = IsZExt ? AArch64::UBFMXri : AArch64::SBFMXri;
|
2014-03-29 18:18:08 +08:00
|
|
|
else
|
2014-09-16 07:20:17 +08:00
|
|
|
Opc = IsZExt ? AArch64::UBFMWri : AArch64::SBFMWri;
|
2014-03-29 18:18:08 +08:00
|
|
|
Imm = 7;
|
|
|
|
break;
|
|
|
|
case MVT::i16:
|
|
|
|
if (DestVT == MVT::i64)
|
2014-09-16 07:20:17 +08:00
|
|
|
Opc = IsZExt ? AArch64::UBFMXri : AArch64::SBFMXri;
|
2014-03-29 18:18:08 +08:00
|
|
|
else
|
2014-09-16 07:20:17 +08:00
|
|
|
Opc = IsZExt ? AArch64::UBFMWri : AArch64::SBFMWri;
|
2014-03-29 18:18:08 +08:00
|
|
|
Imm = 15;
|
|
|
|
break;
|
|
|
|
case MVT::i32:
|
|
|
|
assert(DestVT == MVT::i64 && "IntExt i32 to i32?!?");
|
2014-09-16 07:20:17 +08:00
|
|
|
Opc = IsZExt ? AArch64::UBFMXri : AArch64::SBFMXri;
|
2014-03-29 18:18:08 +08:00
|
|
|
Imm = 31;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Handle i8 and i16 as i32.
|
|
|
|
if (DestVT == MVT::i8 || DestVT == MVT::i16)
|
|
|
|
DestVT = MVT::i32;
|
2014-04-30 17:32:01 +08:00
|
|
|
else if (DestVT == MVT::i64) {
|
2014-05-24 20:50:23 +08:00
|
|
|
unsigned Src64 = MRI.createVirtualRegister(&AArch64::GPR64RegClass);
|
2014-04-30 17:32:01 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
2014-05-24 20:50:23 +08:00
|
|
|
TII.get(AArch64::SUBREG_TO_REG), Src64)
|
2014-04-30 17:32:01 +08:00
|
|
|
.addImm(0)
|
|
|
|
.addReg(SrcReg)
|
2014-05-24 20:50:23 +08:00
|
|
|
.addImm(AArch64::sub_32);
|
2014-04-30 17:32:01 +08:00
|
|
|
SrcReg = Src64;
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-08-22 04:57:57 +08:00
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
(DestVT == MVT::i64) ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
2014-09-04 04:56:59 +08:00
|
|
|
return fastEmitInst_rii(Opc, RC, SrcReg, /*TODO:IsKill=*/false, 0, Imm);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectIntExt(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
// On ARM, in general, integer casts don't involve legal types; this code
|
|
|
|
// handles promotable integers. The high bits for a type smaller than
|
|
|
|
// the register size are assumed to be undefined.
|
|
|
|
Type *DestTy = I->getType();
|
|
|
|
Value *Src = I->getOperand(0);
|
|
|
|
Type *SrcTy = Src->getType();
|
|
|
|
|
|
|
|
unsigned SrcReg = getRegForValue(Src);
|
|
|
|
if (!SrcReg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
EVT SrcEVT = TLI.getValueType(SrcTy, true);
|
|
|
|
EVT DestEVT = TLI.getValueType(DestTy, true);
|
|
|
|
if (!SrcEVT.isSimple())
|
|
|
|
return false;
|
|
|
|
if (!DestEVT.isSimple())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MVT SrcVT = SrcEVT.getSimpleVT();
|
|
|
|
MVT DestVT = DestEVT.getSimpleVT();
|
2014-08-05 13:43:44 +08:00
|
|
|
unsigned ResultReg = 0;
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool IsZExt = isa<ZExtInst>(I);
|
2014-08-05 13:43:44 +08:00
|
|
|
// Check if it is an argument and if it is already zero/sign-extended.
|
|
|
|
if (const auto *Arg = dyn_cast<Argument>(Src)) {
|
2014-09-16 07:20:17 +08:00
|
|
|
if ((IsZExt && Arg->hasZExtAttr()) || (!IsZExt && Arg->hasSExtAttr())) {
|
2014-08-05 15:31:30 +08:00
|
|
|
if (DestVT == MVT::i64) {
|
|
|
|
ResultReg = createResultReg(TLI.getRegClassFor(DestVT));
|
2014-08-05 13:43:44 +08:00
|
|
|
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, DbgLoc,
|
|
|
|
TII.get(AArch64::SUBREG_TO_REG), ResultReg)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(SrcReg)
|
|
|
|
.addImm(AArch64::sub_32);
|
2014-08-05 15:31:30 +08:00
|
|
|
} else
|
|
|
|
ResultReg = SrcReg;
|
2014-08-05 13:43:44 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!ResultReg)
|
2014-09-16 07:20:17 +08:00
|
|
|
ResultReg = emitIntExt(SrcVT, SrcReg, DestVT, IsZExt);
|
2014-08-05 13:43:44 +08:00
|
|
|
|
|
|
|
if (!ResultReg)
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
2014-08-05 13:43:44 +08:00
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectRem(const Instruction *I, unsigned ISDOpcode) {
|
2014-03-29 18:18:08 +08:00
|
|
|
EVT DestEVT = TLI.getValueType(I->getType(), true);
|
|
|
|
if (!DestEVT.isSimple())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MVT DestVT = DestEVT.getSimpleVT();
|
|
|
|
if (DestVT != MVT::i64 && DestVT != MVT::i32)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned DivOpc;
|
2014-09-16 07:20:17 +08:00
|
|
|
bool Is64bit = (DestVT == MVT::i64);
|
2014-03-29 18:18:08 +08:00
|
|
|
switch (ISDOpcode) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case ISD::SREM:
|
2014-09-16 07:20:17 +08:00
|
|
|
DivOpc = Is64bit ? AArch64::SDIVXr : AArch64::SDIVWr;
|
2014-03-29 18:18:08 +08:00
|
|
|
break;
|
|
|
|
case ISD::UREM:
|
2014-09-16 07:20:17 +08:00
|
|
|
DivOpc = Is64bit ? AArch64::UDIVXr : AArch64::UDIVWr;
|
2014-03-29 18:18:08 +08:00
|
|
|
break;
|
|
|
|
}
|
2014-09-16 07:20:17 +08:00
|
|
|
unsigned MSubOpc = Is64bit ? AArch64::MSUBXrrr : AArch64::MSUBWrrr;
|
2014-03-29 18:18:08 +08:00
|
|
|
unsigned Src0Reg = getRegForValue(I->getOperand(0));
|
|
|
|
if (!Src0Reg)
|
|
|
|
return false;
|
2014-08-22 04:57:57 +08:00
|
|
|
bool Src0IsKill = hasTrivialKill(I->getOperand(0));
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
unsigned Src1Reg = getRegForValue(I->getOperand(1));
|
|
|
|
if (!Src1Reg)
|
|
|
|
return false;
|
2014-08-22 04:57:57 +08:00
|
|
|
bool Src1IsKill = hasTrivialKill(I->getOperand(1));
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-08-22 04:57:57 +08:00
|
|
|
const TargetRegisterClass *RC =
|
|
|
|
(DestVT == MVT::i64) ? &AArch64::GPR64RegClass : &AArch64::GPR32RegClass;
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned QuotReg = fastEmitInst_rr(DivOpc, RC, Src0Reg, /*IsKill=*/false,
|
2014-08-22 04:57:57 +08:00
|
|
|
Src1Reg, /*IsKill=*/false);
|
|
|
|
assert(QuotReg && "Unexpected DIV instruction emission failure.");
|
2014-04-17 01:09:20 +08:00
|
|
|
// The remainder is computed as numerator - (quotient * denominator) using the
|
2014-03-29 18:18:08 +08:00
|
|
|
// MSUB instruction.
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned ResultReg = fastEmitInst_rrr(MSubOpc, RC, QuotReg, /*IsKill=*/true,
|
2014-08-22 04:57:57 +08:00
|
|
|
Src1Reg, Src1IsKill, Src0Reg,
|
|
|
|
Src0IsKill);
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectMul(const Instruction *I) {
|
2014-09-18 04:35:41 +08:00
|
|
|
MVT VT;
|
|
|
|
if (!isTypeSupported(I->getType(), VT, /*IsVectorAllowed=*/true))
|
2014-03-29 18:18:08 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-18 04:35:41 +08:00
|
|
|
if (VT.isVector())
|
|
|
|
return selectBinaryOp(I, ISD::MUL);
|
|
|
|
|
|
|
|
const Value *Src0 = I->getOperand(0);
|
|
|
|
const Value *Src1 = I->getOperand(1);
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(Src0))
|
|
|
|
if (C->getValue().isPowerOf2())
|
|
|
|
std::swap(Src0, Src1);
|
|
|
|
|
|
|
|
// Try to simplify to a shift instruction.
|
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(Src1))
|
|
|
|
if (C->getValue().isPowerOf2()) {
|
|
|
|
uint64_t ShiftVal = C->getValue().logBase2();
|
|
|
|
MVT SrcVT = VT;
|
|
|
|
bool IsZExt = true;
|
|
|
|
if (const auto *ZExt = dyn_cast<ZExtInst>(Src0)) {
|
|
|
|
MVT VT;
|
|
|
|
if (isValueAvailable(ZExt) && isTypeSupported(ZExt->getSrcTy(), VT)) {
|
|
|
|
SrcVT = VT;
|
|
|
|
IsZExt = true;
|
|
|
|
Src0 = ZExt->getOperand(0);
|
|
|
|
}
|
|
|
|
} else if (const auto *SExt = dyn_cast<SExtInst>(Src0)) {
|
|
|
|
MVT VT;
|
|
|
|
if (isValueAvailable(SExt) && isTypeSupported(SExt->getSrcTy(), VT)) {
|
|
|
|
SrcVT = VT;
|
|
|
|
IsZExt = false;
|
|
|
|
Src0 = SExt->getOperand(0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned Src0Reg = getRegForValue(Src0);
|
|
|
|
if (!Src0Reg)
|
|
|
|
return false;
|
|
|
|
bool Src0IsKill = hasTrivialKill(Src0);
|
|
|
|
|
|
|
|
unsigned ResultReg =
|
|
|
|
emitLSL_ri(VT, SrcVT, Src0Reg, Src0IsKill, ShiftVal, IsZExt);
|
|
|
|
|
|
|
|
if (ResultReg) {
|
|
|
|
updateValueMap(I, ResultReg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
unsigned Src0Reg = getRegForValue(I->getOperand(0));
|
|
|
|
if (!Src0Reg)
|
|
|
|
return false;
|
2014-07-31 06:04:25 +08:00
|
|
|
bool Src0IsKill = hasTrivialKill(I->getOperand(0));
|
2014-03-29 18:18:08 +08:00
|
|
|
|
|
|
|
unsigned Src1Reg = getRegForValue(I->getOperand(1));
|
|
|
|
if (!Src1Reg)
|
|
|
|
return false;
|
2014-07-31 06:04:25 +08:00
|
|
|
bool Src1IsKill = hasTrivialKill(I->getOperand(1));
|
|
|
|
|
2014-09-18 04:35:41 +08:00
|
|
|
unsigned ResultReg = emitMul_rr(VT, Src0Reg, Src0IsKill, Src1Reg, Src1IsKill);
|
2014-07-31 06:04:25 +08:00
|
|
|
|
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
2014-03-29 18:18:08 +08:00
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-03-29 18:18:08 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectShift(const Instruction *I) {
|
2014-08-27 08:58:26 +08:00
|
|
|
MVT RetVT;
|
2014-09-16 05:27:56 +08:00
|
|
|
if (!isTypeSupported(I->getType(), RetVT, /*IsVectorAllowed=*/true))
|
2014-07-31 06:04:22 +08:00
|
|
|
return false;
|
|
|
|
|
2014-09-16 05:27:56 +08:00
|
|
|
if (RetVT.isVector())
|
|
|
|
return selectOperator(I, I->getOpcode());
|
|
|
|
|
2014-08-22 07:06:07 +08:00
|
|
|
if (const auto *C = dyn_cast<ConstantInt>(I->getOperand(1))) {
|
|
|
|
unsigned ResultReg = 0;
|
|
|
|
uint64_t ShiftVal = C->getZExtValue();
|
2014-08-27 08:58:26 +08:00
|
|
|
MVT SrcVT = RetVT;
|
|
|
|
bool IsZExt = (I->getOpcode() == Instruction::AShr) ? false : true;
|
2014-08-29 08:19:21 +08:00
|
|
|
const Value *Op0 = I->getOperand(0);
|
2014-08-27 08:58:26 +08:00
|
|
|
if (const auto *ZExt = dyn_cast<ZExtInst>(Op0)) {
|
|
|
|
MVT TmpVT;
|
2014-09-03 06:33:53 +08:00
|
|
|
if (isValueAvailable(ZExt) && isTypeSupported(ZExt->getSrcTy(), TmpVT)) {
|
2014-08-27 08:58:26 +08:00
|
|
|
SrcVT = TmpVT;
|
|
|
|
IsZExt = true;
|
|
|
|
Op0 = ZExt->getOperand(0);
|
|
|
|
}
|
|
|
|
} else if (const auto *SExt = dyn_cast<SExtInst>(Op0)) {
|
|
|
|
MVT TmpVT;
|
2014-09-03 06:33:53 +08:00
|
|
|
if (isValueAvailable(SExt) && isTypeSupported(SExt->getSrcTy(), TmpVT)) {
|
2014-08-27 08:58:26 +08:00
|
|
|
SrcVT = TmpVT;
|
|
|
|
IsZExt = false;
|
|
|
|
Op0 = SExt->getOperand(0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned Op0Reg = getRegForValue(Op0);
|
|
|
|
if (!Op0Reg)
|
|
|
|
return false;
|
|
|
|
bool Op0IsKill = hasTrivialKill(Op0);
|
|
|
|
|
2014-08-22 07:06:07 +08:00
|
|
|
switch (I->getOpcode()) {
|
|
|
|
default: llvm_unreachable("Unexpected instruction.");
|
|
|
|
case Instruction::Shl:
|
2014-08-27 08:58:26 +08:00
|
|
|
ResultReg = emitLSL_ri(RetVT, SrcVT, Op0Reg, Op0IsKill, ShiftVal, IsZExt);
|
2014-08-22 07:06:07 +08:00
|
|
|
break;
|
|
|
|
case Instruction::AShr:
|
2014-08-27 08:58:26 +08:00
|
|
|
ResultReg = emitASR_ri(RetVT, SrcVT, Op0Reg, Op0IsKill, ShiftVal, IsZExt);
|
2014-08-22 07:06:07 +08:00
|
|
|
break;
|
|
|
|
case Instruction::LShr:
|
2014-08-27 08:58:26 +08:00
|
|
|
ResultReg = emitLSR_ri(RetVT, SrcVT, Op0Reg, Op0IsKill, ShiftVal, IsZExt);
|
2014-08-22 07:06:07 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
2014-07-31 06:04:22 +08:00
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-08-22 07:06:07 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-08-27 08:58:26 +08:00
|
|
|
unsigned Op0Reg = getRegForValue(I->getOperand(0));
|
|
|
|
if (!Op0Reg)
|
|
|
|
return false;
|
|
|
|
bool Op0IsKill = hasTrivialKill(I->getOperand(0));
|
|
|
|
|
2014-08-22 07:06:07 +08:00
|
|
|
unsigned Op1Reg = getRegForValue(I->getOperand(1));
|
|
|
|
if (!Op1Reg)
|
|
|
|
return false;
|
|
|
|
bool Op1IsKill = hasTrivialKill(I->getOperand(1));
|
|
|
|
|
|
|
|
unsigned ResultReg = 0;
|
|
|
|
switch (I->getOpcode()) {
|
|
|
|
default: llvm_unreachable("Unexpected instruction.");
|
|
|
|
case Instruction::Shl:
|
|
|
|
ResultReg = emitLSL_rr(RetVT, Op0Reg, Op0IsKill, Op1Reg, Op1IsKill);
|
|
|
|
break;
|
|
|
|
case Instruction::AShr:
|
|
|
|
ResultReg = emitASR_rr(RetVT, Op0Reg, Op0IsKill, Op1Reg, Op1IsKill);
|
|
|
|
break;
|
|
|
|
case Instruction::LShr:
|
|
|
|
ResultReg = emitLSR_rr(RetVT, Op0Reg, Op0IsKill, Op1Reg, Op1IsKill);
|
|
|
|
break;
|
2014-07-31 06:04:22 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-07-31 06:04:22 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 07:20:17 +08:00
|
|
|
bool AArch64FastISel::selectBitCast(const Instruction *I) {
|
2014-07-31 14:25:37 +08:00
|
|
|
MVT RetVT, SrcVT;
|
|
|
|
|
|
|
|
if (!isTypeLegal(I->getOperand(0)->getType(), SrcVT))
|
|
|
|
return false;
|
|
|
|
if (!isTypeLegal(I->getType(), RetVT))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned Opc;
|
|
|
|
if (RetVT == MVT::f32 && SrcVT == MVT::i32)
|
|
|
|
Opc = AArch64::FMOVWSr;
|
|
|
|
else if (RetVT == MVT::f64 && SrcVT == MVT::i64)
|
|
|
|
Opc = AArch64::FMOVXDr;
|
|
|
|
else if (RetVT == MVT::i32 && SrcVT == MVT::f32)
|
|
|
|
Opc = AArch64::FMOVSWr;
|
|
|
|
else if (RetVT == MVT::i64 && SrcVT == MVT::f64)
|
|
|
|
Opc = AArch64::FMOVDXr;
|
|
|
|
else
|
|
|
|
return false;
|
|
|
|
|
2014-08-22 04:57:57 +08:00
|
|
|
const TargetRegisterClass *RC = nullptr;
|
|
|
|
switch (RetVT.SimpleTy) {
|
|
|
|
default: llvm_unreachable("Unexpected value type.");
|
|
|
|
case MVT::i32: RC = &AArch64::GPR32RegClass; break;
|
|
|
|
case MVT::i64: RC = &AArch64::GPR64RegClass; break;
|
|
|
|
case MVT::f32: RC = &AArch64::FPR32RegClass; break;
|
|
|
|
case MVT::f64: RC = &AArch64::FPR64RegClass; break;
|
|
|
|
}
|
2014-07-31 14:25:37 +08:00
|
|
|
unsigned Op0Reg = getRegForValue(I->getOperand(0));
|
|
|
|
if (!Op0Reg)
|
|
|
|
return false;
|
|
|
|
bool Op0IsKill = hasTrivialKill(I->getOperand(0));
|
2014-09-04 04:56:59 +08:00
|
|
|
unsigned ResultReg = fastEmitInst_r(Opc, RC, Op0Reg, Op0IsKill);
|
2014-07-31 14:25:37 +08:00
|
|
|
|
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
updateValueMap(I, ResultReg);
|
2014-07-31 14:25:37 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-16 06:07:49 +08:00
|
|
|
bool AArch64FastISel::selectFRem(const Instruction *I) {
|
|
|
|
MVT RetVT;
|
|
|
|
if (!isTypeLegal(I->getType(), RetVT))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
RTLIB::Libcall LC;
|
|
|
|
switch (RetVT.SimpleTy) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case MVT::f32:
|
|
|
|
LC = RTLIB::REM_F32;
|
|
|
|
break;
|
|
|
|
case MVT::f64:
|
|
|
|
LC = RTLIB::REM_F64;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
ArgListTy Args;
|
|
|
|
Args.reserve(I->getNumOperands());
|
|
|
|
|
|
|
|
// Populate the argument list.
|
|
|
|
for (auto &Arg : I->operands()) {
|
|
|
|
ArgListEntry Entry;
|
|
|
|
Entry.Val = Arg;
|
|
|
|
Entry.Ty = Arg->getType();
|
|
|
|
Args.push_back(Entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
CallLoweringInfo CLI;
|
|
|
|
CLI.setCallee(TLI.getLibcallCallingConv(LC), I->getType(),
|
|
|
|
TLI.getLibcallName(LC), std::move(Args));
|
|
|
|
if (!lowerCallTo(CLI))
|
|
|
|
return false;
|
|
|
|
updateValueMap(I, CLI.ResultReg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-18 05:55:55 +08:00
|
|
|
bool AArch64FastISel::selectSDiv(const Instruction *I) {
|
|
|
|
MVT VT;
|
|
|
|
if (!isTypeLegal(I->getType(), VT))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (!isa<ConstantInt>(I->getOperand(1)))
|
|
|
|
return selectBinaryOp(I, ISD::SDIV);
|
|
|
|
|
|
|
|
const APInt &C = cast<ConstantInt>(I->getOperand(1))->getValue();
|
|
|
|
if ((VT != MVT::i32 && VT != MVT::i64) || !C ||
|
|
|
|
!(C.isPowerOf2() || (-C).isPowerOf2()))
|
|
|
|
return selectBinaryOp(I, ISD::SDIV);
|
|
|
|
|
|
|
|
unsigned Lg2 = C.countTrailingZeros();
|
|
|
|
unsigned Src0Reg = getRegForValue(I->getOperand(0));
|
|
|
|
if (!Src0Reg)
|
|
|
|
return false;
|
|
|
|
bool Src0IsKill = hasTrivialKill(I->getOperand(0));
|
|
|
|
|
|
|
|
if (cast<BinaryOperator>(I)->isExact()) {
|
|
|
|
unsigned ResultReg = emitASR_ri(VT, VT, Src0Reg, Src0IsKill, Lg2);
|
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
|
|
|
updateValueMap(I, ResultReg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned Pow2MinusOne = (1 << Lg2) - 1;
|
|
|
|
unsigned AddReg = emitAddSub_ri(/*UseAdd=*/true, VT, Src0Reg,
|
|
|
|
/*IsKill=*/false, Pow2MinusOne);
|
|
|
|
if (!AddReg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// (Src0 < 0) ? Pow2 - 1 : 0;
|
|
|
|
if (!emitICmp_ri(VT, Src0Reg, /*IsKill=*/false, 0))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
unsigned SelectOpc;
|
|
|
|
const TargetRegisterClass *RC;
|
|
|
|
if (VT == MVT::i64) {
|
|
|
|
SelectOpc = AArch64::CSELXr;
|
|
|
|
RC = &AArch64::GPR64RegClass;
|
|
|
|
} else {
|
|
|
|
SelectOpc = AArch64::CSELWr;
|
|
|
|
RC = &AArch64::GPR32RegClass;
|
|
|
|
}
|
|
|
|
unsigned SelectReg =
|
|
|
|
fastEmitInst_rri(SelectOpc, RC, AddReg, /*IsKill=*/true, Src0Reg,
|
|
|
|
Src0IsKill, AArch64CC::LT);
|
|
|
|
if (!SelectReg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Divide by Pow2 --> ashr. If we're dividing by a negative value we must also
|
|
|
|
// negate the result.
|
|
|
|
unsigned ZeroReg = (VT == MVT::i64) ? AArch64::XZR : AArch64::WZR;
|
|
|
|
unsigned ResultReg;
|
|
|
|
if (C.isNegative())
|
|
|
|
ResultReg = emitAddSub_rs(/*UseAdd=*/false, VT, ZeroReg, /*IsKill=*/true,
|
|
|
|
SelectReg, /*IsKill=*/true, AArch64_AM::ASR, Lg2);
|
|
|
|
else
|
|
|
|
ResultReg = emitASR_ri(VT, VT, SelectReg, /*IsKill=*/true, Lg2);
|
|
|
|
|
|
|
|
if (!ResultReg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
updateValueMap(I, ResultReg);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-09-04 04:56:52 +08:00
|
|
|
bool AArch64FastISel::fastSelectInstruction(const Instruction *I) {
|
2014-03-29 18:18:08 +08:00
|
|
|
switch (I->getOpcode()) {
|
|
|
|
default:
|
2014-09-04 09:29:21 +08:00
|
|
|
break;
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::Add:
|
2014-09-03 09:38:36 +08:00
|
|
|
case Instruction::Sub:
|
2014-09-16 05:27:56 +08:00
|
|
|
return selectAddSub(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::Mul:
|
2014-09-18 04:35:41 +08:00
|
|
|
return selectMul(I);
|
2014-09-18 05:55:55 +08:00
|
|
|
case Instruction::SDiv:
|
|
|
|
return selectSDiv(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::SRem:
|
2014-09-04 02:46:45 +08:00
|
|
|
if (!selectBinaryOp(I, ISD::SREM))
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectRem(I, ISD::SREM);
|
2014-09-03 05:32:54 +08:00
|
|
|
return true;
|
|
|
|
case Instruction::URem:
|
2014-09-04 02:46:45 +08:00
|
|
|
if (!selectBinaryOp(I, ISD::UREM))
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectRem(I, ISD::UREM);
|
2014-09-03 05:32:54 +08:00
|
|
|
return true;
|
|
|
|
case Instruction::Shl:
|
|
|
|
case Instruction::LShr:
|
|
|
|
case Instruction::AShr:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectShift(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::And:
|
|
|
|
case Instruction::Or:
|
|
|
|
case Instruction::Xor:
|
2014-09-16 05:27:56 +08:00
|
|
|
return selectLogicalOp(I);
|
2014-09-04 01:58:10 +08:00
|
|
|
case Instruction::Br:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectBranch(I);
|
2014-03-29 18:18:08 +08:00
|
|
|
case Instruction::IndirectBr:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectIndirectBr(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::BitCast:
|
2014-09-04 02:46:45 +08:00
|
|
|
if (!FastISel::selectBitCast(I))
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectBitCast(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
return true;
|
|
|
|
case Instruction::FPToSI:
|
2014-09-04 02:46:45 +08:00
|
|
|
if (!selectCast(I, ISD::FP_TO_SINT))
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectFPToInt(I, /*Signed=*/true);
|
2014-09-03 05:32:54 +08:00
|
|
|
return true;
|
|
|
|
case Instruction::FPToUI:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectFPToInt(I, /*Signed=*/false);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::ZExt:
|
2014-09-04 02:46:45 +08:00
|
|
|
if (!selectCast(I, ISD::ZERO_EXTEND))
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectIntExt(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
return true;
|
|
|
|
case Instruction::SExt:
|
2014-09-04 02:46:45 +08:00
|
|
|
if (!selectCast(I, ISD::SIGN_EXTEND))
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectIntExt(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
return true;
|
|
|
|
case Instruction::Trunc:
|
2014-09-04 02:46:45 +08:00
|
|
|
if (!selectCast(I, ISD::TRUNCATE))
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectTrunc(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
return true;
|
2014-03-29 18:18:08 +08:00
|
|
|
case Instruction::FPExt:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectFPExt(I);
|
2014-03-29 18:18:08 +08:00
|
|
|
case Instruction::FPTrunc:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectFPTrunc(I);
|
2014-03-29 18:18:08 +08:00
|
|
|
case Instruction::SIToFP:
|
2014-09-04 02:46:45 +08:00
|
|
|
if (!selectCast(I, ISD::SINT_TO_FP))
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectIntToFP(I, /*Signed=*/true);
|
2014-09-03 05:32:54 +08:00
|
|
|
return true;
|
2014-03-29 18:18:08 +08:00
|
|
|
case Instruction::UIToFP:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectIntToFP(I, /*Signed=*/false);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::Load:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectLoad(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::Store:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectStore(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::FCmp:
|
|
|
|
case Instruction::ICmp:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectCmp(I);
|
2014-09-03 05:32:54 +08:00
|
|
|
case Instruction::Select:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectSelect(I);
|
2014-03-29 18:18:08 +08:00
|
|
|
case Instruction::Ret:
|
2014-09-16 07:20:17 +08:00
|
|
|
return selectRet(I);
|
2014-09-16 06:07:49 +08:00
|
|
|
case Instruction::FRem:
|
|
|
|
return selectFRem(I);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
2014-09-03 05:32:54 +08:00
|
|
|
|
2014-09-04 09:29:21 +08:00
|
|
|
// fall-back to target-independent instruction selection.
|
|
|
|
return selectOperator(I, I->getOpcode());
|
2014-03-29 18:18:08 +08:00
|
|
|
// Silence warnings.
|
2014-05-24 20:50:23 +08:00
|
|
|
(void)&CC_AArch64_DarwinPCS_VarArg;
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
namespace llvm {
|
2014-09-16 07:20:17 +08:00
|
|
|
llvm::FastISel *AArch64::createFastISel(FunctionLoweringInfo &FuncInfo,
|
|
|
|
const TargetLibraryInfo *LibInfo) {
|
|
|
|
return new AArch64FastISel(FuncInfo, LibInfo);
|
2014-03-29 18:18:08 +08:00
|
|
|
}
|
|
|
|
}
|