2011-12-13 05:14:40 +08:00
|
|
|
//===-- HexagonISelLowering.cpp - Hexagon DAG Lowering Implementation -----===//
|
|
|
|
//
|
|
|
|
// The LLVM Compiler Infrastructure
|
|
|
|
//
|
|
|
|
// This file is distributed under the University of Illinois Open Source
|
|
|
|
// License. See LICENSE.TXT for details.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This file implements the interfaces that Hexagon uses to lower LLVM code
|
|
|
|
// into a selection DAG.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#include "HexagonISelLowering.h"
|
2017-06-06 19:49:48 +08:00
|
|
|
#include "Hexagon.h"
|
2011-12-13 05:14:40 +08:00
|
|
|
#include "HexagonMachineFunctionInfo.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "HexagonRegisterInfo.h"
|
2011-12-13 05:14:40 +08:00
|
|
|
#include "HexagonSubtarget.h"
|
2012-12-04 00:50:05 +08:00
|
|
|
#include "HexagonTargetMachine.h"
|
|
|
|
#include "HexagonTargetObjectFile.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/ADT/APInt.h"
|
|
|
|
#include "llvm/ADT/ArrayRef.h"
|
|
|
|
#include "llvm/ADT/SmallVector.h"
|
2011-12-13 05:14:40 +08:00
|
|
|
#include "llvm/CodeGen/CallingConvLower.h"
|
|
|
|
#include "llvm/CodeGen/MachineFrameInfo.h"
|
|
|
|
#include "llvm/CodeGen/MachineFunction.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/CodeGen/MachineMemOperand.h"
|
2011-12-13 05:14:40 +08:00
|
|
|
#include "llvm/CodeGen/MachineRegisterInfo.h"
|
2017-06-06 19:49:48 +08:00
|
|
|
#include "llvm/CodeGen/RuntimeLibcalls.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/CodeGen/SelectionDAG.h"
|
2017-11-17 09:07:10 +08:00
|
|
|
#include "llvm/CodeGen/TargetCallingConv.h"
|
2011-12-13 05:14:40 +08:00
|
|
|
#include "llvm/CodeGen/ValueTypes.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/IR/BasicBlock.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/CallingConv.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/IR/DataLayout.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/DerivedTypes.h"
|
|
|
|
#include "llvm/IR/Function.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/IR/GlobalValue.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/InlineAsm.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/IR/Instructions.h"
|
2013-01-02 19:36:10 +08:00
|
|
|
#include "llvm/IR/Intrinsics.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/IR/Module.h"
|
|
|
|
#include "llvm/IR/Type.h"
|
|
|
|
#include "llvm/IR/Value.h"
|
|
|
|
#include "llvm/MC/MCRegisterInfo.h"
|
|
|
|
#include "llvm/Support/Casting.h"
|
|
|
|
#include "llvm/Support/CodeGen.h"
|
2012-04-21 23:31:36 +08:00
|
|
|
#include "llvm/Support/CommandLine.h"
|
2011-12-13 05:14:40 +08:00
|
|
|
#include "llvm/Support/Debug.h"
|
|
|
|
#include "llvm/Support/ErrorHandling.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/Support/MathExtras.h"
|
2012-04-21 23:31:45 +08:00
|
|
|
#include "llvm/Support/raw_ostream.h"
|
2016-12-17 09:09:05 +08:00
|
|
|
#include "llvm/Target/TargetMachine.h"
|
|
|
|
#include <algorithm>
|
|
|
|
#include <cassert>
|
|
|
|
#include <cstddef>
|
|
|
|
#include <cstdint>
|
|
|
|
#include <limits>
|
|
|
|
#include <utility>
|
2012-04-21 23:31:36 +08:00
|
|
|
|
2012-03-18 02:46:09 +08:00
|
|
|
using namespace llvm;
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2014-04-22 06:55:11 +08:00
|
|
|
#define DEBUG_TYPE "hexagon-lowering"
|
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
static cl::opt<bool> EmitJumpTables("hexagon-emit-jump-tables",
|
|
|
|
cl::init(true), cl::Hidden,
|
2015-04-23 05:17:00 +08:00
|
|
|
cl::desc("Control jump table emission on Hexagon target"));
|
|
|
|
|
|
|
|
static cl::opt<bool> EnableHexSDNodeSched("enable-hexagon-sdnode-sched",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(false),
|
|
|
|
cl::desc("Enable Hexagon SDNode scheduling"));
|
|
|
|
|
|
|
|
static cl::opt<bool> EnableFastMath("ffast-math",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(false),
|
|
|
|
cl::desc("Enable Fast Math processing"));
|
|
|
|
|
|
|
|
static cl::opt<int> MinimumJumpTables("minimum-jump-tables",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(5),
|
|
|
|
cl::desc("Set minimum jump tables"));
|
|
|
|
|
|
|
|
static cl::opt<int> MaxStoresPerMemcpyCL("max-store-memcpy",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(6),
|
|
|
|
cl::desc("Max #stores to inline memcpy"));
|
|
|
|
|
|
|
|
static cl::opt<int> MaxStoresPerMemcpyOptSizeCL("max-store-memcpy-Os",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(4),
|
|
|
|
cl::desc("Max #stores to inline memcpy"));
|
|
|
|
|
|
|
|
static cl::opt<int> MaxStoresPerMemmoveCL("max-store-memmove",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(6),
|
|
|
|
cl::desc("Max #stores to inline memmove"));
|
|
|
|
|
|
|
|
static cl::opt<int> MaxStoresPerMemmoveOptSizeCL("max-store-memmove-Os",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(4),
|
|
|
|
cl::desc("Max #stores to inline memmove"));
|
|
|
|
|
|
|
|
static cl::opt<int> MaxStoresPerMemsetCL("max-store-memset",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(8),
|
|
|
|
cl::desc("Max #stores to inline memset"));
|
|
|
|
|
|
|
|
static cl::opt<int> MaxStoresPerMemsetOptSizeCL("max-store-memset-Os",
|
|
|
|
cl::Hidden, cl::ZeroOrMore, cl::init(4),
|
|
|
|
cl::desc("Max #stores to inline memset"));
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2013-10-27 19:16:09 +08:00
|
|
|
namespace {
|
2016-12-17 09:09:05 +08:00
|
|
|
|
2016-07-30 00:44:27 +08:00
|
|
|
class HexagonCCState : public CCState {
|
|
|
|
unsigned NumNamedVarArgParams;
|
|
|
|
|
|
|
|
public:
|
|
|
|
HexagonCCState(CallingConv::ID CC, bool isVarArg, MachineFunction &MF,
|
|
|
|
SmallVectorImpl<CCValAssign> &locs, LLVMContext &C,
|
|
|
|
int NumNamedVarArgParams)
|
|
|
|
: CCState(CC, isVarArg, MF, locs, C),
|
|
|
|
NumNamedVarArgParams(NumNamedVarArgParams) {}
|
|
|
|
|
|
|
|
unsigned getNumNamedVarArgParams() const { return NumNamedVarArgParams; }
|
|
|
|
};
|
|
|
|
|
|
|
|
enum StridedLoadKind {
|
|
|
|
Even = 0,
|
|
|
|
Odd,
|
|
|
|
NoPattern
|
|
|
|
};
|
2016-12-17 09:09:05 +08:00
|
|
|
|
|
|
|
} // end anonymous namespace
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Implement calling convention for Hexagon.
|
2015-11-27 02:38:27 +08:00
|
|
|
|
2017-12-15 02:35:24 +08:00
|
|
|
static const MVT LegalV64[] = { MVT::v64i8, MVT::v32i16, MVT::v16i32 };
|
|
|
|
static const MVT LegalW64[] = { MVT::v128i8, MVT::v64i16, MVT::v32i32 };
|
|
|
|
static const MVT LegalV128[] = { MVT::v128i8, MVT::v64i16, MVT::v32i32 };
|
|
|
|
static const MVT LegalW128[] = { MVT::v256i8, MVT::v128i16, MVT::v64i32 };
|
2017-12-07 00:40:37 +08:00
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
static bool
|
|
|
|
CC_Hexagon(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State);
|
|
|
|
|
|
|
|
static bool
|
|
|
|
CC_Hexagon32(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State);
|
|
|
|
|
|
|
|
static bool
|
|
|
|
CC_Hexagon64(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State);
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
static bool
|
|
|
|
CC_HexagonVector(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State);
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
static bool
|
|
|
|
RetCC_Hexagon(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State);
|
|
|
|
|
|
|
|
static bool
|
|
|
|
RetCC_Hexagon32(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State);
|
|
|
|
|
|
|
|
static bool
|
|
|
|
RetCC_Hexagon64(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State);
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
static bool
|
|
|
|
RetCC_HexagonVector(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State);
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
static bool
|
|
|
|
CC_Hexagon_VarArg (unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
2013-10-27 19:16:09 +08:00
|
|
|
HexagonCCState &HState = static_cast<HexagonCCState &>(State);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
if (ValNo < HState.getNumNamedVarArgParams()) {
|
2011-12-13 05:14:40 +08:00
|
|
|
// Deal with named arguments.
|
|
|
|
return CC_Hexagon(ValNo, ValVT, LocVT, LocInfo, ArgFlags, State);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Deal with un-named arguments.
|
2016-08-02 04:31:50 +08:00
|
|
|
unsigned Offset;
|
2011-12-13 05:14:40 +08:00
|
|
|
if (ArgFlags.isByVal()) {
|
|
|
|
// If pass-by-value, the size allocated on stack is decided
|
|
|
|
// by ArgFlags.getByValSize(), not by the size of LocVT.
|
2016-08-02 04:31:50 +08:00
|
|
|
Offset = State.AllocateStack(ArgFlags.getByValSize(),
|
|
|
|
ArgFlags.getByValAlign());
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
2011-12-13 05:14:40 +08:00
|
|
|
return false;
|
|
|
|
}
|
2013-03-08 04:28:34 +08:00
|
|
|
if (LocVT == MVT::i1 || LocVT == MVT::i8 || LocVT == MVT::i16) {
|
|
|
|
LocVT = MVT::i32;
|
|
|
|
ValVT = MVT::i32;
|
|
|
|
if (ArgFlags.isSExt())
|
|
|
|
LocInfo = CCValAssign::SExt;
|
|
|
|
else if (ArgFlags.isZExt())
|
|
|
|
LocInfo = CCValAssign::ZExt;
|
|
|
|
else
|
|
|
|
LocInfo = CCValAssign::AExt;
|
|
|
|
}
|
2012-05-11 04:20:25 +08:00
|
|
|
if (LocVT == MVT::i32 || LocVT == MVT::f32) {
|
2016-08-02 04:31:50 +08:00
|
|
|
Offset = State.AllocateStack(4, 4);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
2011-12-13 05:14:40 +08:00
|
|
|
return false;
|
|
|
|
}
|
2012-05-11 04:20:25 +08:00
|
|
|
if (LocVT == MVT::i64 || LocVT == MVT::f64) {
|
2016-08-02 04:31:50 +08:00
|
|
|
Offset = State.AllocateStack(8, 8);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
2011-12-13 05:14:40 +08:00
|
|
|
return false;
|
|
|
|
}
|
2015-11-27 02:38:27 +08:00
|
|
|
if (LocVT == MVT::v2i64 || LocVT == MVT::v4i32 || LocVT == MVT::v8i16 ||
|
|
|
|
LocVT == MVT::v16i8) {
|
2016-08-02 04:31:50 +08:00
|
|
|
Offset = State.AllocateStack(16, 16);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
2015-11-27 02:38:27 +08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
if (LocVT == MVT::v4i64 || LocVT == MVT::v8i32 || LocVT == MVT::v16i16 ||
|
|
|
|
LocVT == MVT::v32i8) {
|
2016-08-02 04:31:50 +08:00
|
|
|
Offset = State.AllocateStack(32, 32);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
2015-11-27 02:38:27 +08:00
|
|
|
return false;
|
|
|
|
}
|
2017-12-15 02:35:24 +08:00
|
|
|
if (LocVT == MVT::v16i32 || LocVT == MVT::v32i16 ||
|
2015-11-27 02:38:27 +08:00
|
|
|
LocVT == MVT::v64i8 || LocVT == MVT::v512i1) {
|
2016-08-02 04:31:50 +08:00
|
|
|
Offset = State.AllocateStack(64, 64);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
2015-11-27 02:38:27 +08:00
|
|
|
return false;
|
|
|
|
}
|
2017-12-15 02:35:24 +08:00
|
|
|
if (LocVT == MVT::v32i32 || LocVT == MVT::v64i16 ||
|
2015-11-27 02:38:27 +08:00
|
|
|
LocVT == MVT::v128i8 || LocVT == MVT::v1024i1) {
|
2016-08-02 04:31:50 +08:00
|
|
|
Offset = State.AllocateStack(128, 128);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
2015-11-27 02:38:27 +08:00
|
|
|
return false;
|
|
|
|
}
|
2017-12-15 02:35:24 +08:00
|
|
|
if (LocVT == MVT::v64i32 || LocVT == MVT::v128i16 ||
|
2015-11-27 02:38:27 +08:00
|
|
|
LocVT == MVT::v256i8) {
|
2016-08-02 04:31:50 +08:00
|
|
|
Offset = State.AllocateStack(256, 256);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
2015-11-27 02:38:27 +08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2014-04-28 12:05:08 +08:00
|
|
|
llvm_unreachable(nullptr);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
static bool CC_Hexagon (unsigned ValNo, MVT ValVT, MVT LocVT,
|
|
|
|
CCValAssign::LocInfo LocInfo, ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
2011-12-13 05:14:40 +08:00
|
|
|
if (ArgFlags.isByVal()) {
|
|
|
|
// Passed on stack.
|
2015-04-23 05:17:00 +08:00
|
|
|
unsigned Offset = State.AllocateStack(ArgFlags.getByValSize(),
|
|
|
|
ArgFlags.getByValAlign());
|
2011-12-13 05:14:40 +08:00
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2017-03-02 01:30:10 +08:00
|
|
|
if (LocVT == MVT::i1) {
|
|
|
|
LocVT = MVT::i32;
|
|
|
|
} else if (LocVT == MVT::i8 || LocVT == MVT::i16) {
|
2011-12-13 05:14:40 +08:00
|
|
|
LocVT = MVT::i32;
|
|
|
|
ValVT = MVT::i32;
|
|
|
|
if (ArgFlags.isSExt())
|
|
|
|
LocInfo = CCValAssign::SExt;
|
|
|
|
else if (ArgFlags.isZExt())
|
|
|
|
LocInfo = CCValAssign::ZExt;
|
|
|
|
else
|
|
|
|
LocInfo = CCValAssign::AExt;
|
2015-03-20 00:33:08 +08:00
|
|
|
} else if (LocVT == MVT::v4i8 || LocVT == MVT::v2i16) {
|
|
|
|
LocVT = MVT::i32;
|
|
|
|
LocInfo = CCValAssign::BCvt;
|
|
|
|
} else if (LocVT == MVT::v8i8 || LocVT == MVT::v4i16 || LocVT == MVT::v2i32) {
|
|
|
|
LocVT = MVT::i64;
|
|
|
|
LocInfo = CCValAssign::BCvt;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2012-05-11 04:20:25 +08:00
|
|
|
if (LocVT == MVT::i32 || LocVT == MVT::f32) {
|
2011-12-13 05:14:40 +08:00
|
|
|
if (!CC_Hexagon32(ValNo, ValVT, LocVT, LocInfo, ArgFlags, State))
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2012-05-11 04:20:25 +08:00
|
|
|
if (LocVT == MVT::i64 || LocVT == MVT::f64) {
|
2011-12-13 05:14:40 +08:00
|
|
|
if (!CC_Hexagon64(ValNo, ValVT, LocVT, LocInfo, ArgFlags, State))
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
if (LocVT == MVT::v8i32 || LocVT == MVT::v16i16 || LocVT == MVT::v32i8) {
|
|
|
|
unsigned Offset = State.AllocateStack(ArgFlags.getByValSize(), 32);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2017-11-28 02:12:16 +08:00
|
|
|
auto &HST = State.getMachineFunction().getSubtarget<HexagonSubtarget>();
|
|
|
|
if (HST.isHVXVectorType(LocVT)) {
|
2015-11-27 02:38:27 +08:00
|
|
|
if (!CC_HexagonVector(ValNo, ValVT, LocVT, LocInfo, ArgFlags, State))
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
return true; // CC didn't match.
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static bool CC_Hexagon32(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
2014-04-04 13:16:06 +08:00
|
|
|
static const MCPhysReg RegList[] = {
|
2011-12-13 05:14:40 +08:00
|
|
|
Hexagon::R0, Hexagon::R1, Hexagon::R2, Hexagon::R3, Hexagon::R4,
|
|
|
|
Hexagon::R5
|
|
|
|
};
|
2015-02-21 10:11:17 +08:00
|
|
|
if (unsigned Reg = State.AllocateReg(RegList)) {
|
2011-12-13 05:14:40 +08:00
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned Offset = State.AllocateStack(4, 4);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool CC_Hexagon64(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
|
|
|
if (unsigned Reg = State.AllocateReg(Hexagon::D0)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2014-04-04 13:16:06 +08:00
|
|
|
static const MCPhysReg RegList1[] = {
|
2011-12-13 05:14:40 +08:00
|
|
|
Hexagon::D1, Hexagon::D2
|
|
|
|
};
|
2014-04-04 13:16:06 +08:00
|
|
|
static const MCPhysReg RegList2[] = {
|
2011-12-13 05:14:40 +08:00
|
|
|
Hexagon::R1, Hexagon::R3
|
|
|
|
};
|
2015-02-21 10:11:17 +08:00
|
|
|
if (unsigned Reg = State.AllocateReg(RegList1, RegList2)) {
|
2011-12-13 05:14:40 +08:00
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned Offset = State.AllocateStack(8, 8, Hexagon::D2);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
static bool CC_HexagonVector(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
2016-08-02 04:31:50 +08:00
|
|
|
static const MCPhysReg VecLstS[] = {
|
|
|
|
Hexagon::V0, Hexagon::V1, Hexagon::V2, Hexagon::V3, Hexagon::V4,
|
|
|
|
Hexagon::V5, Hexagon::V6, Hexagon::V7, Hexagon::V8, Hexagon::V9,
|
|
|
|
Hexagon::V10, Hexagon::V11, Hexagon::V12, Hexagon::V13, Hexagon::V14,
|
|
|
|
Hexagon::V15
|
|
|
|
};
|
|
|
|
static const MCPhysReg VecLstD[] = {
|
|
|
|
Hexagon::W0, Hexagon::W1, Hexagon::W2, Hexagon::W3, Hexagon::W4,
|
|
|
|
Hexagon::W5, Hexagon::W6, Hexagon::W7
|
|
|
|
};
|
2015-11-27 02:38:27 +08:00
|
|
|
auto &MF = State.getMachineFunction();
|
|
|
|
auto &HST = MF.getSubtarget<HexagonSubtarget>();
|
|
|
|
|
2017-10-19 02:07:07 +08:00
|
|
|
if (HST.useHVX64BOps() &&
|
2017-12-15 02:35:24 +08:00
|
|
|
(LocVT == MVT::v16i32 || LocVT == MVT::v32i16 ||
|
2015-11-27 02:38:27 +08:00
|
|
|
LocVT == MVT::v64i8 || LocVT == MVT::v512i1)) {
|
|
|
|
if (unsigned Reg = State.AllocateReg(VecLstS)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
unsigned Offset = State.AllocateStack(64, 64);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
2017-12-15 02:35:24 +08:00
|
|
|
if (HST.useHVX64BOps() && (LocVT == MVT::v32i32 ||
|
2017-10-19 01:45:22 +08:00
|
|
|
LocVT == MVT::v64i16 || LocVT == MVT::v128i8)) {
|
2015-11-27 02:38:27 +08:00
|
|
|
if (unsigned Reg = State.AllocateReg(VecLstD)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
unsigned Offset = State.AllocateStack(128, 128);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
2017-10-19 01:45:22 +08:00
|
|
|
// 128B Mode
|
2017-12-15 02:35:24 +08:00
|
|
|
if (HST.useHVX128BOps() && (LocVT == MVT::v64i32 ||
|
2017-10-19 02:07:07 +08:00
|
|
|
LocVT == MVT::v128i16 || LocVT == MVT::v256i8)) {
|
2015-11-27 02:38:27 +08:00
|
|
|
if (unsigned Reg = State.AllocateReg(VecLstD)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
unsigned Offset = State.AllocateStack(256, 256);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
2017-10-19 02:07:07 +08:00
|
|
|
if (HST.useHVX128BOps() &&
|
2017-12-15 02:35:24 +08:00
|
|
|
(LocVT == MVT::v32i32 || LocVT == MVT::v64i16 ||
|
2015-11-27 02:38:27 +08:00
|
|
|
LocVT == MVT::v128i8 || LocVT == MVT::v1024i1)) {
|
|
|
|
if (unsigned Reg = State.AllocateReg(VecLstS)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
unsigned Offset = State.AllocateStack(128, 128);
|
|
|
|
State.addLoc(CCValAssign::getMem(ValNo, ValVT, Offset, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
static bool RetCC_Hexagon(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
2015-11-27 02:38:27 +08:00
|
|
|
auto &MF = State.getMachineFunction();
|
|
|
|
auto &HST = MF.getSubtarget<HexagonSubtarget>();
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2016-03-05 01:38:05 +08:00
|
|
|
if (LocVT == MVT::i1) {
|
|
|
|
// Return values of type MVT::i1 still need to be assigned to R0, but
|
|
|
|
// the value type needs to remain i1. LowerCallResult will deal with it,
|
|
|
|
// but it needs to recognize i1 as the value type.
|
|
|
|
LocVT = MVT::i32;
|
|
|
|
} else if (LocVT == MVT::i8 || LocVT == MVT::i16) {
|
2011-12-13 05:14:40 +08:00
|
|
|
LocVT = MVT::i32;
|
|
|
|
ValVT = MVT::i32;
|
|
|
|
if (ArgFlags.isSExt())
|
|
|
|
LocInfo = CCValAssign::SExt;
|
|
|
|
else if (ArgFlags.isZExt())
|
|
|
|
LocInfo = CCValAssign::ZExt;
|
|
|
|
else
|
|
|
|
LocInfo = CCValAssign::AExt;
|
2015-03-20 00:33:08 +08:00
|
|
|
} else if (LocVT == MVT::v4i8 || LocVT == MVT::v2i16) {
|
|
|
|
LocVT = MVT::i32;
|
|
|
|
LocInfo = CCValAssign::BCvt;
|
|
|
|
} else if (LocVT == MVT::v8i8 || LocVT == MVT::v4i16 || LocVT == MVT::v2i32) {
|
|
|
|
LocVT = MVT::i64;
|
|
|
|
LocInfo = CCValAssign::BCvt;
|
2015-11-27 02:38:27 +08:00
|
|
|
} else if (LocVT == MVT::v64i8 || LocVT == MVT::v32i16 ||
|
2017-12-15 02:35:24 +08:00
|
|
|
LocVT == MVT::v16i32 || LocVT == MVT::v512i1) {
|
2015-11-27 02:38:27 +08:00
|
|
|
LocVT = MVT::v16i32;
|
|
|
|
ValVT = MVT::v16i32;
|
|
|
|
LocInfo = CCValAssign::Full;
|
|
|
|
} else if (LocVT == MVT::v128i8 || LocVT == MVT::v64i16 ||
|
2017-12-15 02:35:24 +08:00
|
|
|
LocVT == MVT::v32i32 ||
|
2017-10-19 02:07:07 +08:00
|
|
|
(LocVT == MVT::v1024i1 && HST.useHVX128BOps())) {
|
2015-11-27 02:38:27 +08:00
|
|
|
LocVT = MVT::v32i32;
|
|
|
|
ValVT = MVT::v32i32;
|
|
|
|
LocInfo = CCValAssign::Full;
|
|
|
|
} else if (LocVT == MVT::v256i8 || LocVT == MVT::v128i16 ||
|
2017-12-15 02:35:24 +08:00
|
|
|
LocVT == MVT::v64i32) {
|
2015-11-27 02:38:27 +08:00
|
|
|
LocVT = MVT::v64i32;
|
|
|
|
ValVT = MVT::v64i32;
|
|
|
|
LocInfo = CCValAssign::Full;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
2012-05-11 04:20:25 +08:00
|
|
|
if (LocVT == MVT::i32 || LocVT == MVT::f32) {
|
2011-12-13 05:14:40 +08:00
|
|
|
if (!RetCC_Hexagon32(ValNo, ValVT, LocVT, LocInfo, ArgFlags, State))
|
2016-08-02 04:31:50 +08:00
|
|
|
return false;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2012-05-11 04:20:25 +08:00
|
|
|
if (LocVT == MVT::i64 || LocVT == MVT::f64) {
|
2011-12-13 05:14:40 +08:00
|
|
|
if (!RetCC_Hexagon64(ValNo, ValVT, LocVT, LocInfo, ArgFlags, State))
|
2016-08-02 04:31:50 +08:00
|
|
|
return false;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
2015-11-27 02:38:27 +08:00
|
|
|
if (LocVT == MVT::v16i32 || LocVT == MVT::v32i32 || LocVT == MVT::v64i32) {
|
|
|
|
if (!RetCC_HexagonVector(ValNo, ValVT, LocVT, LocInfo, ArgFlags, State))
|
2016-08-02 04:31:50 +08:00
|
|
|
return false;
|
2015-11-27 02:38:27 +08:00
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
return true; // CC didn't match.
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool RetCC_Hexagon32(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
2012-05-11 04:20:25 +08:00
|
|
|
if (LocVT == MVT::i32 || LocVT == MVT::f32) {
|
2016-07-19 01:36:46 +08:00
|
|
|
// Note that use of registers beyond R1 is not ABI compliant. However there
|
|
|
|
// are (experimental) IR passes which generate internal functions that
|
|
|
|
// return structs using these additional registers.
|
|
|
|
static const uint16_t RegList[] = { Hexagon::R0, Hexagon::R1,
|
|
|
|
Hexagon::R2, Hexagon::R3,
|
2016-08-02 04:31:50 +08:00
|
|
|
Hexagon::R4, Hexagon::R5 };
|
2016-07-19 01:36:46 +08:00
|
|
|
if (unsigned Reg = State.AllocateReg(RegList)) {
|
2011-12-13 05:14:40 +08:00
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-04-13 23:05:51 +08:00
|
|
|
return true;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static bool RetCC_Hexagon64(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
2012-05-11 04:20:25 +08:00
|
|
|
if (LocVT == MVT::i64 || LocVT == MVT::f64) {
|
2011-12-13 05:14:40 +08:00
|
|
|
if (unsigned Reg = State.AllocateReg(Hexagon::D0)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-04-13 23:05:51 +08:00
|
|
|
return true;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
static bool RetCC_HexagonVector(unsigned ValNo, MVT ValVT,
|
|
|
|
MVT LocVT, CCValAssign::LocInfo LocInfo,
|
|
|
|
ISD::ArgFlagsTy ArgFlags, CCState &State) {
|
|
|
|
auto &MF = State.getMachineFunction();
|
|
|
|
auto &HST = MF.getSubtarget<HexagonSubtarget>();
|
|
|
|
|
|
|
|
if (LocVT == MVT::v16i32) {
|
|
|
|
if (unsigned Reg = State.AllocateReg(Hexagon::V0)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
} else if (LocVT == MVT::v32i32) {
|
2017-10-19 02:07:07 +08:00
|
|
|
unsigned Req = HST.useHVX128BOps() ? Hexagon::V0 : Hexagon::W0;
|
2015-11-27 02:38:27 +08:00
|
|
|
if (unsigned Reg = State.AllocateReg(Req)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
} else if (LocVT == MVT::v64i32) {
|
|
|
|
if (unsigned Reg = State.AllocateReg(Hexagon::W0)) {
|
|
|
|
State.addLoc(CCValAssign::getReg(ValNo, ValVT, Reg, LocVT, LocInfo));
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-04-13 23:05:51 +08:00
|
|
|
return true;
|
2015-11-27 02:38:27 +08:00
|
|
|
}
|
|
|
|
|
2016-04-15 14:20:21 +08:00
|
|
|
void HexagonTargetLowering::promoteLdStType(MVT VT, MVT PromotedLdStVT) {
|
2015-12-19 04:19:30 +08:00
|
|
|
if (VT != PromotedLdStVT) {
|
2016-04-15 14:20:21 +08:00
|
|
|
setOperationAction(ISD::LOAD, VT, Promote);
|
|
|
|
AddPromotedToType(ISD::LOAD, VT, PromotedLdStVT);
|
2015-12-19 04:19:30 +08:00
|
|
|
|
2016-04-15 14:20:21 +08:00
|
|
|
setOperationAction(ISD::STORE, VT, Promote);
|
|
|
|
AddPromotedToType(ISD::STORE, VT, PromotedLdStVT);
|
2015-12-19 04:19:30 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG)
|
2016-08-02 04:31:50 +08:00
|
|
|
const {
|
2011-12-13 05:14:40 +08:00
|
|
|
return SDValue();
|
|
|
|
}
|
|
|
|
|
|
|
|
/// CreateCopyOfByValArgument - Make a copy of an aggregate at address specified
|
|
|
|
/// by "Src" to address "Dst" of size "Size". Alignment information is
|
|
|
|
/// specified by the specific parameter attribute. The copy will be passed as
|
|
|
|
/// a byval function parameter. Sometimes what we are copying is the end of a
|
|
|
|
/// larger object, the part that does not fit in registers.
|
2016-06-12 23:39:02 +08:00
|
|
|
static SDValue CreateCopyOfByValArgument(SDValue Src, SDValue Dst,
|
|
|
|
SDValue Chain, ISD::ArgFlagsTy Flags,
|
|
|
|
SelectionDAG &DAG, const SDLoc &dl) {
|
2015-04-28 22:05:47 +08:00
|
|
|
SDValue SizeNode = DAG.getConstant(Flags.getByValSize(), dl, MVT::i32);
|
2011-12-13 05:14:40 +08:00
|
|
|
return DAG.getMemcpy(Chain, dl, Dst, Src, SizeNode, Flags.getByValAlign(),
|
|
|
|
/*isVolatile=*/false, /*AlwaysInline=*/false,
|
2015-04-14 01:16:45 +08:00
|
|
|
/*isTailCall=*/false,
|
2011-12-13 05:14:40 +08:00
|
|
|
MachinePointerInfo(), MachinePointerInfo());
|
|
|
|
}
|
|
|
|
|
2017-04-13 23:05:51 +08:00
|
|
|
bool
|
|
|
|
HexagonTargetLowering::CanLowerReturn(
|
|
|
|
CallingConv::ID CallConv, MachineFunction &MF, bool isVarArg,
|
|
|
|
const SmallVectorImpl<ISD::OutputArg> &Outs,
|
|
|
|
LLVMContext &Context) const {
|
|
|
|
SmallVector<CCValAssign, 16> RVLocs;
|
|
|
|
CCState CCInfo(CallConv, isVarArg, MF, RVLocs, Context);
|
|
|
|
return CCInfo.CheckReturn(Outs, RetCC_Hexagon);
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
// LowerReturn - Lower ISD::RET. If a struct is larger than 8 bytes and is
|
|
|
|
// passed by value, the function prototype is modified to return void and
|
|
|
|
// the value is stored in memory pointed by a pointer passed by caller.
|
|
|
|
SDValue
|
2016-06-12 23:39:02 +08:00
|
|
|
HexagonTargetLowering::LowerReturn(SDValue Chain, CallingConv::ID CallConv,
|
|
|
|
bool isVarArg,
|
2011-12-13 05:14:40 +08:00
|
|
|
const SmallVectorImpl<ISD::OutputArg> &Outs,
|
|
|
|
const SmallVectorImpl<SDValue> &OutVals,
|
2016-06-12 23:39:02 +08:00
|
|
|
const SDLoc &dl, SelectionDAG &DAG) const {
|
2011-12-13 05:14:40 +08:00
|
|
|
// CCValAssign - represent the assignment of the return value to locations.
|
|
|
|
SmallVector<CCValAssign, 16> RVLocs;
|
|
|
|
|
|
|
|
// CCState - Info about the registers and stack slot.
|
2014-08-07 02:45:26 +08:00
|
|
|
CCState CCInfo(CallConv, isVarArg, DAG.getMachineFunction(), RVLocs,
|
|
|
|
*DAG.getContext());
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Analyze return values of ISD::RET
|
|
|
|
CCInfo.AnalyzeReturn(Outs, RetCC_Hexagon);
|
|
|
|
|
|
|
|
SDValue Flag;
|
2013-02-06 02:08:43 +08:00
|
|
|
SmallVector<SDValue, 4> RetOps(1, Chain);
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
// Copy the result values into the output registers.
|
|
|
|
for (unsigned i = 0; i != RVLocs.size(); ++i) {
|
|
|
|
CCValAssign &VA = RVLocs[i];
|
|
|
|
|
|
|
|
Chain = DAG.getCopyToReg(Chain, dl, VA.getLocReg(), OutVals[i], Flag);
|
|
|
|
|
|
|
|
// Guarantee that all emitted copies are stuck together with flags.
|
|
|
|
Flag = Chain.getValue(1);
|
2013-02-06 02:08:43 +08:00
|
|
|
RetOps.push_back(DAG.getRegister(VA.getLocReg(), VA.getLocVT()));
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2013-02-06 02:08:43 +08:00
|
|
|
RetOps[0] = Chain; // Update chain.
|
|
|
|
|
|
|
|
// Add the flag if we have it.
|
2011-12-13 05:14:40 +08:00
|
|
|
if (Flag.getNode())
|
2013-02-06 02:08:43 +08:00
|
|
|
RetOps.push_back(Flag);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2014-04-27 02:35:24 +08:00
|
|
|
return DAG.getNode(HexagonISD::RET_FLAG, dl, MVT::Other, RetOps);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2017-04-19 05:16:46 +08:00
|
|
|
bool HexagonTargetLowering::mayBeEmittedAsTailCall(const CallInst *CI) const {
|
2015-04-23 05:17:00 +08:00
|
|
|
// If either no tail call or told not to tail call at all, don't.
|
2015-06-10 03:07:19 +08:00
|
|
|
auto Attr =
|
|
|
|
CI->getParent()->getParent()->getFnAttribute("disable-tail-calls");
|
|
|
|
if (!CI->isTailCall() || Attr.getValueAsString() == "true")
|
2015-04-23 05:17:00 +08:00
|
|
|
return false;
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
return true;
|
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
/// LowerCallResult - Lower the result values of an ISD::CALL into the
|
|
|
|
/// appropriate copies out of appropriate physical registers. This assumes that
|
2017-02-18 06:14:51 +08:00
|
|
|
/// Chain/Glue are the input chain/glue to use, and that TheCall is the call
|
2011-12-13 05:14:40 +08:00
|
|
|
/// being lowered. Returns a SDNode with the same number of values as the
|
|
|
|
/// ISD::CALL.
|
2016-06-12 23:39:02 +08:00
|
|
|
SDValue HexagonTargetLowering::LowerCallResult(
|
2017-02-18 06:14:51 +08:00
|
|
|
SDValue Chain, SDValue Glue, CallingConv::ID CallConv, bool isVarArg,
|
2016-06-12 23:39:02 +08:00
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins, const SDLoc &dl,
|
|
|
|
SelectionDAG &DAG, SmallVectorImpl<SDValue> &InVals,
|
|
|
|
const SmallVectorImpl<SDValue> &OutVals, SDValue Callee) const {
|
2011-12-13 05:14:40 +08:00
|
|
|
// Assign locations to each value returned by this call.
|
|
|
|
SmallVector<CCValAssign, 16> RVLocs;
|
|
|
|
|
2014-08-07 02:45:26 +08:00
|
|
|
CCState CCInfo(CallConv, isVarArg, DAG.getMachineFunction(), RVLocs,
|
|
|
|
*DAG.getContext());
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
CCInfo.AnalyzeCallResult(Ins, RetCC_Hexagon);
|
|
|
|
|
|
|
|
// Copy all of the result registers out of their specified physreg.
|
|
|
|
for (unsigned i = 0; i != RVLocs.size(); ++i) {
|
2016-03-05 01:38:05 +08:00
|
|
|
SDValue RetVal;
|
|
|
|
if (RVLocs[i].getValVT() == MVT::i1) {
|
|
|
|
// Return values of type MVT::i1 require special handling. The reason
|
|
|
|
// is that MVT::i1 is associated with the PredRegs register class, but
|
|
|
|
// values of that type are still returned in R0. Generate an explicit
|
|
|
|
// copy into a predicate register from R0, and treat the value of the
|
|
|
|
// predicate register as the call result.
|
|
|
|
auto &MRI = DAG.getMachineFunction().getRegInfo();
|
|
|
|
SDValue FR0 = DAG.getCopyFromReg(Chain, dl, RVLocs[i].getLocReg(),
|
2017-02-18 06:14:51 +08:00
|
|
|
MVT::i32, Glue);
|
2016-03-05 01:38:05 +08:00
|
|
|
// FR0 = (Value, Chain, Glue)
|
|
|
|
unsigned PredR = MRI.createVirtualRegister(&Hexagon::PredRegsRegClass);
|
|
|
|
SDValue TPR = DAG.getCopyToReg(FR0.getValue(1), dl, PredR,
|
|
|
|
FR0.getValue(0), FR0.getValue(2));
|
|
|
|
// TPR = (Chain, Glue)
|
2017-02-18 06:14:51 +08:00
|
|
|
// Don't glue this CopyFromReg, because it copies from a virtual
|
|
|
|
// register. If it is glued to the call, InstrEmitter will add it
|
|
|
|
// as an implicit def to the call (EmitMachineNode).
|
|
|
|
RetVal = DAG.getCopyFromReg(TPR.getValue(0), dl, PredR, MVT::i1);
|
|
|
|
Glue = TPR.getValue(1);
|
2017-10-24 03:35:25 +08:00
|
|
|
Chain = TPR.getValue(0);
|
2016-03-05 01:38:05 +08:00
|
|
|
} else {
|
|
|
|
RetVal = DAG.getCopyFromReg(Chain, dl, RVLocs[i].getLocReg(),
|
2017-02-18 06:14:51 +08:00
|
|
|
RVLocs[i].getValVT(), Glue);
|
|
|
|
Glue = RetVal.getValue(2);
|
2017-10-24 03:35:25 +08:00
|
|
|
Chain = RetVal.getValue(1);
|
2016-03-05 01:38:05 +08:00
|
|
|
}
|
|
|
|
InVals.push_back(RetVal.getValue(0));
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return Chain;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// LowerCall - Functions arguments are copied from virtual regs to
|
|
|
|
/// (physical regs)/(stack frame), CALLSEQ_START and CALLSEQ_END are emitted.
|
|
|
|
SDValue
|
2012-05-26 00:35:28 +08:00
|
|
|
HexagonTargetLowering::LowerCall(TargetLowering::CallLoweringInfo &CLI,
|
2011-12-13 05:14:40 +08:00
|
|
|
SmallVectorImpl<SDValue> &InVals) const {
|
2012-05-26 00:35:28 +08:00
|
|
|
SelectionDAG &DAG = CLI.DAG;
|
2013-07-14 12:42:23 +08:00
|
|
|
SDLoc &dl = CLI.DL;
|
|
|
|
SmallVectorImpl<ISD::OutputArg> &Outs = CLI.Outs;
|
|
|
|
SmallVectorImpl<SDValue> &OutVals = CLI.OutVals;
|
|
|
|
SmallVectorImpl<ISD::InputArg> &Ins = CLI.Ins;
|
2012-05-26 00:35:28 +08:00
|
|
|
SDValue Chain = CLI.Chain;
|
|
|
|
SDValue Callee = CLI.Callee;
|
2016-08-02 04:31:50 +08:00
|
|
|
bool &IsTailCall = CLI.IsTailCall;
|
2012-05-26 00:35:28 +08:00
|
|
|
CallingConv::ID CallConv = CLI.CallConv;
|
2016-08-02 04:31:50 +08:00
|
|
|
bool IsVarArg = CLI.IsVarArg;
|
|
|
|
bool DoesNotReturn = CLI.DoesNotReturn;
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
bool IsStructRet = (Outs.empty()) ? false : Outs[0].Flags.isSRet();
|
2015-04-23 00:43:53 +08:00
|
|
|
MachineFunction &MF = DAG.getMachineFunction();
|
2017-07-12 01:11:54 +08:00
|
|
|
MachineFrameInfo &MFI = MF.getFrameInfo();
|
2015-07-09 10:09:04 +08:00
|
|
|
auto PtrVT = getPointerTy(MF.getDataLayout());
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Check for varargs.
|
2016-08-02 04:31:50 +08:00
|
|
|
unsigned NumNamedVarArgParams = -1U;
|
2015-12-19 04:19:30 +08:00
|
|
|
if (GlobalAddressSDNode *GAN = dyn_cast<GlobalAddressSDNode>(Callee)) {
|
|
|
|
const GlobalValue *GV = GAN->getGlobal();
|
|
|
|
Callee = DAG.getTargetGlobalAddress(GV, dl, MVT::i32);
|
|
|
|
if (const Function* F = dyn_cast<Function>(GV)) {
|
2011-12-13 05:14:40 +08:00
|
|
|
// If a function has zero args and is a vararg function, that's
|
|
|
|
// disallowed so it must be an undeclared function. Do not assume
|
|
|
|
// varargs if the callee is undefined.
|
2015-12-19 04:19:30 +08:00
|
|
|
if (F->isVarArg() && F->getFunctionType()->getNumParams() != 0)
|
|
|
|
NumNamedVarArgParams = F->getFunctionType()->getNumParams();
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-10-27 19:16:09 +08:00
|
|
|
// Analyze operands of the call, assigning locations to each operand.
|
|
|
|
SmallVector<CCValAssign, 16> ArgLocs;
|
2016-08-02 04:31:50 +08:00
|
|
|
HexagonCCState CCInfo(CallConv, IsVarArg, DAG.getMachineFunction(), ArgLocs,
|
2014-08-07 02:45:26 +08:00
|
|
|
*DAG.getContext(), NumNamedVarArgParams);
|
2013-10-27 19:16:09 +08:00
|
|
|
|
2016-08-02 04:31:50 +08:00
|
|
|
if (IsVarArg)
|
2011-12-13 05:14:40 +08:00
|
|
|
CCInfo.AnalyzeCallOperands(Outs, CC_Hexagon_VarArg);
|
|
|
|
else
|
|
|
|
CCInfo.AnalyzeCallOperands(Outs, CC_Hexagon);
|
|
|
|
|
2017-12-16 06:22:58 +08:00
|
|
|
auto Attr = MF.getFunction().getFnAttribute("disable-tail-calls");
|
2015-06-10 03:07:19 +08:00
|
|
|
if (Attr.getValueAsString() == "true")
|
2016-08-02 04:31:50 +08:00
|
|
|
IsTailCall = false;
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2016-08-02 04:31:50 +08:00
|
|
|
if (IsTailCall) {
|
2017-12-16 06:22:58 +08:00
|
|
|
bool StructAttrFlag = MF.getFunction().hasStructRetAttr();
|
2016-08-02 04:31:50 +08:00
|
|
|
IsTailCall = IsEligibleForTailCallOptimization(Callee, CallConv,
|
|
|
|
IsVarArg, IsStructRet,
|
2011-12-13 05:14:40 +08:00
|
|
|
StructAttrFlag,
|
|
|
|
Outs, OutVals, Ins, DAG);
|
2015-04-23 00:43:53 +08:00
|
|
|
for (unsigned i = 0, e = ArgLocs.size(); i != e; ++i) {
|
2011-12-13 05:14:40 +08:00
|
|
|
CCValAssign &VA = ArgLocs[i];
|
|
|
|
if (VA.isMemLoc()) {
|
2016-08-02 04:31:50 +08:00
|
|
|
IsTailCall = false;
|
2011-12-13 05:14:40 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2016-08-02 04:31:50 +08:00
|
|
|
DEBUG(dbgs() << (IsTailCall ? "Eligible for Tail Call\n"
|
2015-04-23 00:43:53 +08:00
|
|
|
: "Argument must be passed on stack. "
|
|
|
|
"Not eligible for Tail Call\n"));
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
// Get a count of how many bytes are to be pushed on the stack.
|
|
|
|
unsigned NumBytes = CCInfo.getNextStackOffset();
|
|
|
|
SmallVector<std::pair<unsigned, SDValue>, 16> RegsToPass;
|
|
|
|
SmallVector<SDValue, 8> MemOpChains;
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
auto &HRI = *Subtarget.getRegisterInfo();
|
2015-07-09 10:09:04 +08:00
|
|
|
SDValue StackPtr =
|
|
|
|
DAG.getCopyFromReg(Chain, dl, HRI.getStackRegister(), PtrVT);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
bool NeedsArgAlign = false;
|
|
|
|
unsigned LargestAlignSeen = 0;
|
2011-12-13 05:14:40 +08:00
|
|
|
// Walk the register/memloc assignments, inserting copies/loads.
|
|
|
|
for (unsigned i = 0, e = ArgLocs.size(); i != e; ++i) {
|
|
|
|
CCValAssign &VA = ArgLocs[i];
|
|
|
|
SDValue Arg = OutVals[i];
|
|
|
|
ISD::ArgFlagsTy Flags = Outs[i].Flags;
|
2015-11-27 02:38:27 +08:00
|
|
|
// Record if we need > 8 byte alignment on an argument.
|
2017-11-28 02:12:16 +08:00
|
|
|
bool ArgAlign = Subtarget.isHVXVectorType(VA.getValVT());
|
2015-11-27 02:38:27 +08:00
|
|
|
NeedsArgAlign |= ArgAlign;
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Promote the value if needed.
|
|
|
|
switch (VA.getLocInfo()) {
|
|
|
|
default:
|
|
|
|
// Loc info must be one of Full, SExt, ZExt, or AExt.
|
2012-02-07 10:50:20 +08:00
|
|
|
llvm_unreachable("Unknown loc info!");
|
2015-04-23 00:43:53 +08:00
|
|
|
case CCValAssign::BCvt:
|
2011-12-13 05:14:40 +08:00
|
|
|
case CCValAssign::Full:
|
|
|
|
break;
|
|
|
|
case CCValAssign::SExt:
|
|
|
|
Arg = DAG.getNode(ISD::SIGN_EXTEND, dl, VA.getLocVT(), Arg);
|
|
|
|
break;
|
|
|
|
case CCValAssign::ZExt:
|
|
|
|
Arg = DAG.getNode(ISD::ZERO_EXTEND, dl, VA.getLocVT(), Arg);
|
|
|
|
break;
|
|
|
|
case CCValAssign::AExt:
|
|
|
|
Arg = DAG.getNode(ISD::ANY_EXTEND, dl, VA.getLocVT(), Arg);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (VA.isMemLoc()) {
|
|
|
|
unsigned LocMemOffset = VA.getLocMemOffset();
|
2015-04-28 22:05:47 +08:00
|
|
|
SDValue MemAddr = DAG.getConstant(LocMemOffset, dl,
|
|
|
|
StackPtr.getValueType());
|
2015-04-23 00:43:53 +08:00
|
|
|
MemAddr = DAG.getNode(ISD::ADD, dl, MVT::i32, StackPtr, MemAddr);
|
2015-11-27 02:38:27 +08:00
|
|
|
if (ArgAlign)
|
|
|
|
LargestAlignSeen = std::max(LargestAlignSeen,
|
|
|
|
VA.getLocVT().getStoreSizeInBits() >> 3);
|
2011-12-13 05:14:40 +08:00
|
|
|
if (Flags.isByVal()) {
|
|
|
|
// The argument is a struct passed by value. According to LLVM, "Arg"
|
|
|
|
// is is pointer.
|
2015-04-23 00:43:53 +08:00
|
|
|
MemOpChains.push_back(CreateCopyOfByValArgument(Arg, MemAddr, Chain,
|
2011-12-13 05:14:40 +08:00
|
|
|
Flags, DAG, dl));
|
|
|
|
} else {
|
2015-08-12 07:09:45 +08:00
|
|
|
MachinePointerInfo LocPI = MachinePointerInfo::getStack(
|
|
|
|
DAG.getMachineFunction(), LocMemOffset);
|
[SelectionDAG] Get rid of bool parameters in SelectionDAG::getLoad, getStore, and friends.
Summary:
Instead, we take a single flags arg (a bitset).
Also add a default 0 alignment, and change the order of arguments so the
alignment comes before the flags.
This greatly simplifies many callsites, and fixes a bug in
AMDGPUISelLowering, wherein the order of the args to getLoad was
inverted. It also greatly simplifies the process of adding another flag
to getLoad.
Reviewers: chandlerc, tstellarAMD
Subscribers: jholewinski, arsenm, jyknight, dsanders, nemanjai, llvm-commits
Differential Revision: http://reviews.llvm.org/D22249
llvm-svn: 275592
2016-07-16 02:27:10 +08:00
|
|
|
SDValue S = DAG.getStore(Chain, dl, Arg, MemAddr, LocPI);
|
2015-04-23 00:43:53 +08:00
|
|
|
MemOpChains.push_back(S);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Arguments that can be passed on register must be kept at RegsToPass
|
|
|
|
// vector.
|
2015-04-23 00:43:53 +08:00
|
|
|
if (VA.isRegLoc())
|
2011-12-13 05:14:40 +08:00
|
|
|
RegsToPass.push_back(std::make_pair(VA.getLocReg(), Arg));
|
|
|
|
}
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
if (NeedsArgAlign && Subtarget.hasV60TOps()) {
|
|
|
|
DEBUG(dbgs() << "Function needs byte stack align due to call args\n");
|
|
|
|
// V6 vectors passed by value have 64 or 128 byte alignment depending
|
|
|
|
// on whether we are 64 byte vector mode or 128 byte.
|
2017-10-19 02:07:07 +08:00
|
|
|
bool UseHVX128B = Subtarget.useHVX128BOps();
|
2015-11-27 02:38:27 +08:00
|
|
|
assert(Subtarget.useHVXOps());
|
2017-10-19 02:07:07 +08:00
|
|
|
const unsigned ObjAlign = UseHVX128B ? 128 : 64;
|
2015-11-27 02:38:27 +08:00
|
|
|
LargestAlignSeen = std::max(LargestAlignSeen, ObjAlign);
|
2016-07-29 02:40:00 +08:00
|
|
|
MFI.ensureMaxAlignment(LargestAlignSeen);
|
2015-11-27 02:38:27 +08:00
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
// Transform all store nodes into one single node because all store
|
|
|
|
// nodes are independent of each other.
|
2015-04-23 00:43:53 +08:00
|
|
|
if (!MemOpChains.empty())
|
2014-04-27 02:35:24 +08:00
|
|
|
Chain = DAG.getNode(ISD::TokenFactor, dl, MVT::Other, MemOpChains);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2017-02-18 06:14:51 +08:00
|
|
|
SDValue Glue;
|
2016-08-02 04:31:50 +08:00
|
|
|
if (!IsTailCall) {
|
2017-05-09 21:35:13 +08:00
|
|
|
Chain = DAG.getCALLSEQ_START(Chain, NumBytes, 0, dl);
|
2017-02-18 06:14:51 +08:00
|
|
|
Glue = Chain.getValue(1);
|
2015-04-23 00:43:53 +08:00
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Build a sequence of copy-to-reg nodes chained together with token
|
|
|
|
// chain and flag operands which copy the outgoing args into registers.
|
2016-08-02 04:31:50 +08:00
|
|
|
// The Glue is necessary since all emitted instructions must be
|
2011-12-13 05:14:40 +08:00
|
|
|
// stuck together.
|
2016-08-02 04:31:50 +08:00
|
|
|
if (!IsTailCall) {
|
2011-12-13 05:14:40 +08:00
|
|
|
for (unsigned i = 0, e = RegsToPass.size(); i != e; ++i) {
|
|
|
|
Chain = DAG.getCopyToReg(Chain, dl, RegsToPass[i].first,
|
2016-08-02 04:31:50 +08:00
|
|
|
RegsToPass[i].second, Glue);
|
|
|
|
Glue = Chain.getValue(1);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
2015-04-23 00:43:53 +08:00
|
|
|
} else {
|
|
|
|
// For tail calls lower the arguments to the 'real' stack slot.
|
|
|
|
//
|
2011-12-13 05:14:40 +08:00
|
|
|
// Force all the incoming stack arguments to be loaded from the stack
|
|
|
|
// before any new outgoing arguments are stored to the stack, because the
|
|
|
|
// outgoing stack slots may alias the incoming argument stack slots, and
|
|
|
|
// the alias isn't otherwise explicit. This is slightly more conservative
|
|
|
|
// than necessary, because it means that each store effectively depends
|
|
|
|
// on every argument instead of just those arguments it would clobber.
|
|
|
|
//
|
2012-06-02 18:20:22 +08:00
|
|
|
// Do not flag preceding copytoreg stuff together with the following stuff.
|
2016-08-02 04:31:50 +08:00
|
|
|
Glue = SDValue();
|
2011-12-13 05:14:40 +08:00
|
|
|
for (unsigned i = 0, e = RegsToPass.size(); i != e; ++i) {
|
|
|
|
Chain = DAG.getCopyToReg(Chain, dl, RegsToPass[i].first,
|
2016-08-02 04:31:50 +08:00
|
|
|
RegsToPass[i].second, Glue);
|
|
|
|
Glue = Chain.getValue(1);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
2016-08-02 04:31:50 +08:00
|
|
|
Glue = SDValue();
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2016-07-25 22:42:11 +08:00
|
|
|
bool LongCalls = MF.getSubtarget<HexagonSubtarget>().useLongCalls();
|
|
|
|
unsigned Flags = LongCalls ? HexagonII::HMOTF_ConstExtended : 0;
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
// If the callee is a GlobalAddress/ExternalSymbol node (quite common, every
|
|
|
|
// direct call is) turn it into a TargetGlobalAddress/TargetExternalSymbol
|
|
|
|
// node so that legalize doesn't hack it.
|
2015-12-17 01:29:37 +08:00
|
|
|
if (GlobalAddressSDNode *G = dyn_cast<GlobalAddressSDNode>(Callee)) {
|
2016-07-25 22:42:11 +08:00
|
|
|
Callee = DAG.getTargetGlobalAddress(G->getGlobal(), dl, PtrVT, 0, Flags);
|
2011-12-13 05:14:40 +08:00
|
|
|
} else if (ExternalSymbolSDNode *S =
|
|
|
|
dyn_cast<ExternalSymbolSDNode>(Callee)) {
|
2016-07-25 22:42:11 +08:00
|
|
|
Callee = DAG.getTargetExternalSymbol(S->getSymbol(), PtrVT, Flags);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Returns a chain & a flag for retval copy to use.
|
|
|
|
SDVTList NodeTys = DAG.getVTList(MVT::Other, MVT::Glue);
|
|
|
|
SmallVector<SDValue, 8> Ops;
|
|
|
|
Ops.push_back(Chain);
|
|
|
|
Ops.push_back(Callee);
|
|
|
|
|
|
|
|
// Add argument registers to the end of the list so that they are
|
|
|
|
// known live into the call.
|
|
|
|
for (unsigned i = 0, e = RegsToPass.size(); i != e; ++i) {
|
|
|
|
Ops.push_back(DAG.getRegister(RegsToPass[i].first,
|
|
|
|
RegsToPass[i].second.getValueType()));
|
|
|
|
}
|
|
|
|
|
2017-02-18 06:14:51 +08:00
|
|
|
const uint32_t *Mask = HRI.getCallPreservedMask(MF, CallConv);
|
|
|
|
assert(Mask && "Missing call preserved mask for calling convention");
|
|
|
|
Ops.push_back(DAG.getRegisterMask(Mask));
|
|
|
|
|
2016-08-02 04:31:50 +08:00
|
|
|
if (Glue.getNode())
|
|
|
|
Ops.push_back(Glue);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2016-08-02 04:31:50 +08:00
|
|
|
if (IsTailCall) {
|
2017-07-12 01:11:54 +08:00
|
|
|
MFI.setHasTailCall();
|
2014-04-27 02:35:24 +08:00
|
|
|
return DAG.getNode(HexagonISD::TC_RETURN, dl, NodeTys, Ops);
|
2015-05-09 07:52:00 +08:00
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2017-07-12 01:11:54 +08:00
|
|
|
// Set this here because we need to know this for "hasFP" in frame lowering.
|
|
|
|
// The target-independent code calls getFrameRegister before setting it, and
|
|
|
|
// getFrameRegister uses hasFP to determine whether the function has FP.
|
|
|
|
MFI.setHasCalls(true);
|
|
|
|
|
2016-08-12 19:12:02 +08:00
|
|
|
unsigned OpCode = DoesNotReturn ? HexagonISD::CALLnr : HexagonISD::CALL;
|
2015-01-17 01:05:27 +08:00
|
|
|
Chain = DAG.getNode(OpCode, dl, NodeTys, Ops);
|
2016-08-02 04:31:50 +08:00
|
|
|
Glue = Chain.getValue(1);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Create the CALLSEQ_END node.
|
2015-04-28 22:05:47 +08:00
|
|
|
Chain = DAG.getCALLSEQ_END(Chain, DAG.getIntPtrConstant(NumBytes, dl, true),
|
2016-08-02 04:31:50 +08:00
|
|
|
DAG.getIntPtrConstant(0, dl, true), Glue, dl);
|
|
|
|
Glue = Chain.getValue(1);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Handle result values, copying them out of physregs into vregs that we
|
|
|
|
// return.
|
2016-08-02 04:31:50 +08:00
|
|
|
return LowerCallResult(Chain, Glue, CallConv, IsVarArg, Ins, dl, DAG,
|
2011-12-13 05:14:40 +08:00
|
|
|
InVals, OutVals, Callee);
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool getIndexedAddressParts(SDNode *Ptr, EVT VT,
|
2016-08-02 04:31:50 +08:00
|
|
|
SDValue &Base, SDValue &Offset,
|
|
|
|
bool &IsInc, SelectionDAG &DAG) {
|
2011-12-13 05:14:40 +08:00
|
|
|
if (Ptr->getOpcode() != ISD::ADD)
|
2015-04-23 00:43:53 +08:00
|
|
|
return false;
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
auto &HST = static_cast<const HexagonSubtarget&>(DAG.getSubtarget());
|
|
|
|
|
2017-10-19 02:07:07 +08:00
|
|
|
bool ValidHVX128BType =
|
2017-12-15 02:35:24 +08:00
|
|
|
HST.useHVX128BOps() && (VT == MVT::v32i32 ||
|
2017-10-19 02:07:07 +08:00
|
|
|
VT == MVT::v64i16 || VT == MVT::v128i8);
|
2015-11-27 02:38:27 +08:00
|
|
|
bool ValidHVXType =
|
2017-12-15 02:35:24 +08:00
|
|
|
HST.useHVX64BOps() && (VT == MVT::v16i32 ||
|
2015-11-27 02:38:27 +08:00
|
|
|
VT == MVT::v32i16 || VT == MVT::v64i8);
|
|
|
|
|
2017-10-19 02:07:07 +08:00
|
|
|
if (ValidHVX128BType || ValidHVXType || VT == MVT::i64 || VT == MVT::i32 ||
|
|
|
|
VT == MVT::i16 || VT == MVT::i8) {
|
2016-08-02 04:31:50 +08:00
|
|
|
IsInc = (Ptr->getOpcode() == ISD::ADD);
|
2011-12-13 05:14:40 +08:00
|
|
|
Base = Ptr->getOperand(0);
|
|
|
|
Offset = Ptr->getOperand(1);
|
|
|
|
// Ensure that Offset is a constant.
|
2016-08-02 04:31:50 +08:00
|
|
|
return isa<ConstantSDNode>(Offset);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// getPostIndexedAddressParts - returns true by value, base pointer and
|
|
|
|
/// offset pointer and addressing mode by reference if this node can be
|
|
|
|
/// combined with a load / store to form a post-indexed load / store.
|
|
|
|
bool HexagonTargetLowering::getPostIndexedAddressParts(SDNode *N, SDNode *Op,
|
|
|
|
SDValue &Base,
|
|
|
|
SDValue &Offset,
|
|
|
|
ISD::MemIndexedMode &AM,
|
|
|
|
SelectionDAG &DAG) const
|
|
|
|
{
|
|
|
|
EVT VT;
|
|
|
|
|
|
|
|
if (LoadSDNode *LD = dyn_cast<LoadSDNode>(N)) {
|
|
|
|
VT = LD->getMemoryVT();
|
|
|
|
} else if (StoreSDNode *ST = dyn_cast<StoreSDNode>(N)) {
|
|
|
|
VT = ST->getMemoryVT();
|
2016-08-02 04:31:50 +08:00
|
|
|
if (ST->getValue().getValueType() == MVT::i64 && ST->isTruncatingStore())
|
2011-12-13 05:14:40 +08:00
|
|
|
return false;
|
|
|
|
} else {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2016-08-02 04:31:50 +08:00
|
|
|
bool IsInc = false;
|
|
|
|
bool isLegal = getIndexedAddressParts(Op, VT, Base, Offset, IsInc, DAG);
|
2015-11-27 02:38:27 +08:00
|
|
|
if (isLegal) {
|
|
|
|
auto &HII = *Subtarget.getInstrInfo();
|
|
|
|
int32_t OffsetVal = cast<ConstantSDNode>(Offset.getNode())->getSExtValue();
|
|
|
|
if (HII.isValidAutoIncImm(VT, OffsetVal)) {
|
2016-08-02 04:31:50 +08:00
|
|
|
AM = IsInc ? ISD::POST_INC : ISD::POST_DEC;
|
2015-11-27 02:38:27 +08:00
|
|
|
return true;
|
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerINLINEASM(SDValue Op, SelectionDAG &DAG) const {
|
2011-12-13 05:14:40 +08:00
|
|
|
MachineFunction &MF = DAG.getMachineFunction();
|
2017-07-01 05:21:40 +08:00
|
|
|
auto &HMFI = *MF.getInfo<HexagonMachineFunctionInfo>();
|
|
|
|
const HexagonRegisterInfo &HRI = *Subtarget.getRegisterInfo();
|
|
|
|
unsigned LR = HRI.getRARegister();
|
|
|
|
|
|
|
|
if (Op.getOpcode() != ISD::INLINEASM || HMFI.hasClobberLR())
|
|
|
|
return Op;
|
|
|
|
|
|
|
|
unsigned NumOps = Op.getNumOperands();
|
|
|
|
if (Op.getOperand(NumOps-1).getValueType() == MVT::Glue)
|
|
|
|
--NumOps; // Ignore the flag operand.
|
|
|
|
|
|
|
|
for (unsigned i = InlineAsm::Op_FirstOperand; i != NumOps;) {
|
|
|
|
unsigned Flags = cast<ConstantSDNode>(Op.getOperand(i))->getZExtValue();
|
|
|
|
unsigned NumVals = InlineAsm::getNumOperandRegisters(Flags);
|
|
|
|
++i; // Skip the ID value.
|
|
|
|
|
|
|
|
switch (InlineAsm::getKind(Flags)) {
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Bad flags!");
|
|
|
|
case InlineAsm::Kind_RegUse:
|
|
|
|
case InlineAsm::Kind_Imm:
|
|
|
|
case InlineAsm::Kind_Mem:
|
|
|
|
i += NumVals;
|
|
|
|
break;
|
|
|
|
case InlineAsm::Kind_Clobber:
|
|
|
|
case InlineAsm::Kind_RegDef:
|
|
|
|
case InlineAsm::Kind_RegDefEarlyClobber: {
|
|
|
|
for (; NumVals; --NumVals, ++i) {
|
|
|
|
unsigned Reg = cast<RegisterSDNode>(Op.getOperand(i))->getReg();
|
|
|
|
if (Reg != LR)
|
|
|
|
continue;
|
|
|
|
HMFI.setHasClobberLR(true);
|
|
|
|
return Op;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
2017-07-01 05:21:40 +08:00
|
|
|
break;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
}
|
2017-07-01 05:21:40 +08:00
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
return Op;
|
|
|
|
}
|
|
|
|
|
2016-02-18 21:58:38 +08:00
|
|
|
// Need to transform ISD::PREFETCH into something that doesn't inherit
|
|
|
|
// all of the properties of ISD::PREFETCH, specifically SDNPMayLoad and
|
|
|
|
// SDNPMayStore.
|
|
|
|
SDValue HexagonTargetLowering::LowerPREFETCH(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
SDValue Chain = Op.getOperand(0);
|
|
|
|
SDValue Addr = Op.getOperand(1);
|
|
|
|
// Lower it to DCFETCH($reg, #0). A "pat" will try to merge the offset in,
|
|
|
|
// if the "reg" is fed by an "add".
|
|
|
|
SDLoc DL(Op);
|
|
|
|
SDValue Zero = DAG.getConstant(0, DL, MVT::i32);
|
|
|
|
return DAG.getNode(HexagonISD::DCFETCH, DL, MVT::Other, Chain, Addr, Zero);
|
|
|
|
}
|
|
|
|
|
2017-02-23 06:28:47 +08:00
|
|
|
// Custom-handle ISD::READCYCLECOUNTER because the target-independent SDNode
|
|
|
|
// is marked as having side-effects, while the register read on Hexagon does
|
|
|
|
// not have any. TableGen refuses to accept the direct pattern from that node
|
|
|
|
// to the A4_tfrcpp.
|
|
|
|
SDValue HexagonTargetLowering::LowerREADCYCLECOUNTER(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
SDValue Chain = Op.getOperand(0);
|
|
|
|
SDLoc dl(Op);
|
|
|
|
SDVTList VTs = DAG.getVTList(MVT::i32, MVT::Other);
|
|
|
|
return DAG.getNode(HexagonISD::READCYCLE, dl, VTs, Chain);
|
|
|
|
}
|
|
|
|
|
2016-02-18 21:58:38 +08:00
|
|
|
SDValue HexagonTargetLowering::LowerINTRINSIC_VOID(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
SDValue Chain = Op.getOperand(0);
|
|
|
|
unsigned IntNo = cast<ConstantSDNode>(Op.getOperand(1))->getZExtValue();
|
|
|
|
// Lower the hexagon_prefetch builtin to DCFETCH, as above.
|
|
|
|
if (IntNo == Intrinsic::hexagon_prefetch) {
|
|
|
|
SDValue Addr = Op.getOperand(2);
|
|
|
|
SDLoc DL(Op);
|
|
|
|
SDValue Zero = DAG.getConstant(0, DL, MVT::i32);
|
|
|
|
return DAG.getNode(HexagonISD::DCFETCH, DL, MVT::Other, Chain, Addr, Zero);
|
|
|
|
}
|
|
|
|
return SDValue();
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerDYNAMIC_STACKALLOC(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
SDValue Chain = Op.getOperand(0);
|
|
|
|
SDValue Size = Op.getOperand(1);
|
2015-04-23 00:43:53 +08:00
|
|
|
SDValue Align = Op.getOperand(2);
|
2013-05-25 10:42:55 +08:00
|
|
|
SDLoc dl(Op);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-04-23 00:43:53 +08:00
|
|
|
ConstantSDNode *AlignConst = dyn_cast<ConstantSDNode>(Align);
|
|
|
|
assert(AlignConst && "Non-constant Align in LowerDYNAMIC_STACKALLOC");
|
|
|
|
|
|
|
|
unsigned A = AlignConst->getSExtValue();
|
2015-04-23 05:17:00 +08:00
|
|
|
auto &HFI = *Subtarget.getFrameLowering();
|
2015-04-23 00:43:53 +08:00
|
|
|
// "Zero" means natural stack alignment.
|
|
|
|
if (A == 0)
|
|
|
|
A = HFI.getStackAlignment();
|
|
|
|
|
|
|
|
DEBUG({
|
2016-10-20 08:22:23 +08:00
|
|
|
dbgs () << __func__ << " Align: " << A << " Size: ";
|
2015-04-23 00:43:53 +08:00
|
|
|
Size.getNode()->dump(&DAG);
|
|
|
|
dbgs() << "\n";
|
|
|
|
});
|
|
|
|
|
2015-04-28 22:05:47 +08:00
|
|
|
SDValue AC = DAG.getConstant(A, dl, MVT::i32);
|
2015-04-23 00:43:53 +08:00
|
|
|
SDVTList VTs = DAG.getVTList(MVT::i32, MVT::Other);
|
2015-10-20 02:30:27 +08:00
|
|
|
SDValue AA = DAG.getNode(HexagonISD::ALLOCA, dl, VTs, Chain, Size, AC);
|
2016-06-24 01:52:57 +08:00
|
|
|
|
|
|
|
DAG.ReplaceAllUsesOfValueWith(Op, AA);
|
2015-10-20 02:30:27 +08:00
|
|
|
return AA;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2016-06-12 23:39:02 +08:00
|
|
|
SDValue HexagonTargetLowering::LowerFormalArguments(
|
|
|
|
SDValue Chain, CallingConv::ID CallConv, bool isVarArg,
|
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins, const SDLoc &dl,
|
|
|
|
SelectionDAG &DAG, SmallVectorImpl<SDValue> &InVals) const {
|
2011-12-13 05:14:40 +08:00
|
|
|
MachineFunction &MF = DAG.getMachineFunction();
|
2016-07-29 02:40:00 +08:00
|
|
|
MachineFrameInfo &MFI = MF.getFrameInfo();
|
2011-12-13 05:14:40 +08:00
|
|
|
MachineRegisterInfo &RegInfo = MF.getRegInfo();
|
2015-04-23 00:43:53 +08:00
|
|
|
auto &FuncInfo = *MF.getInfo<HexagonMachineFunctionInfo>();
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Assign locations to all of the incoming arguments.
|
|
|
|
SmallVector<CCValAssign, 16> ArgLocs;
|
2014-08-07 02:45:26 +08:00
|
|
|
CCState CCInfo(CallConv, isVarArg, DAG.getMachineFunction(), ArgLocs,
|
|
|
|
*DAG.getContext());
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
CCInfo.AnalyzeFormalArguments(Ins, CC_Hexagon);
|
|
|
|
|
|
|
|
// For LLVM, in the case when returning a struct by value (>8byte),
|
|
|
|
// the first argument is a pointer that points to the location on caller's
|
|
|
|
// stack where the return value will be stored. For Hexagon, the location on
|
|
|
|
// caller's stack is passed only when the struct size is smaller than (and
|
|
|
|
// equal to) 8 bytes. If not, no address will be passed into callee and
|
|
|
|
// callee return the result direclty through R0/R1.
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
SmallVector<SDValue, 8> MemOps;
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
for (unsigned i = 0, e = ArgLocs.size(); i != e; ++i) {
|
|
|
|
CCValAssign &VA = ArgLocs[i];
|
|
|
|
ISD::ArgFlagsTy Flags = Ins[i].Flags;
|
|
|
|
unsigned ObjSize;
|
|
|
|
unsigned StackLocation;
|
|
|
|
int FI;
|
|
|
|
|
|
|
|
if ( (VA.isRegLoc() && !Flags.isByVal())
|
|
|
|
|| (VA.isRegLoc() && Flags.isByVal() && Flags.getByValSize() > 8)) {
|
|
|
|
// Arguments passed in registers
|
|
|
|
// 1. int, long long, ptr args that get allocated in register.
|
|
|
|
// 2. Large struct that gets an register to put its address in.
|
|
|
|
EVT RegVT = VA.getLocVT();
|
2012-05-11 04:20:25 +08:00
|
|
|
if (RegVT == MVT::i8 || RegVT == MVT::i16 ||
|
|
|
|
RegVT == MVT::i32 || RegVT == MVT::f32) {
|
2017-11-23 04:43:00 +08:00
|
|
|
unsigned VReg =
|
2012-04-20 15:30:17 +08:00
|
|
|
RegInfo.createVirtualRegister(&Hexagon::IntRegsRegClass);
|
2011-12-13 05:14:40 +08:00
|
|
|
RegInfo.addLiveIn(VA.getLocReg(), VReg);
|
2017-03-02 01:30:10 +08:00
|
|
|
SDValue Copy = DAG.getCopyFromReg(Chain, dl, VReg, RegVT);
|
|
|
|
// Treat values of type MVT::i1 specially: they are passed in
|
|
|
|
// registers of type i32, but they need to remain as values of
|
|
|
|
// type i1 for consistency of the argument lowering.
|
|
|
|
if (VA.getValVT() == MVT::i1) {
|
|
|
|
// Generate a copy into a predicate register and use the value
|
|
|
|
// of the register as the "InVal".
|
|
|
|
unsigned PReg =
|
|
|
|
RegInfo.createVirtualRegister(&Hexagon::PredRegsRegClass);
|
|
|
|
SDNode *T = DAG.getMachineNode(Hexagon::C2_tfrrp, dl, MVT::i1,
|
|
|
|
Copy.getValue(0));
|
|
|
|
Copy = DAG.getCopyToReg(Copy.getValue(1), dl, PReg, SDValue(T, 0));
|
|
|
|
Copy = DAG.getCopyFromReg(Copy, dl, PReg, MVT::i1);
|
|
|
|
}
|
|
|
|
InVals.push_back(Copy);
|
|
|
|
Chain = Copy.getValue(1);
|
2015-01-29 06:08:16 +08:00
|
|
|
} else if (RegVT == MVT::i64 || RegVT == MVT::f64) {
|
2011-12-13 05:14:40 +08:00
|
|
|
unsigned VReg =
|
2012-04-20 15:30:17 +08:00
|
|
|
RegInfo.createVirtualRegister(&Hexagon::DoubleRegsRegClass);
|
2011-12-13 05:14:40 +08:00
|
|
|
RegInfo.addLiveIn(VA.getLocReg(), VReg);
|
|
|
|
InVals.push_back(DAG.getCopyFromReg(Chain, dl, VReg, RegVT));
|
2015-11-27 02:38:27 +08:00
|
|
|
|
|
|
|
// Single Vector
|
2017-12-15 02:35:24 +08:00
|
|
|
} else if ((RegVT == MVT::v16i32 ||
|
2015-11-27 02:38:27 +08:00
|
|
|
RegVT == MVT::v32i16 || RegVT == MVT::v64i8)) {
|
|
|
|
unsigned VReg =
|
2017-09-15 23:46:05 +08:00
|
|
|
RegInfo.createVirtualRegister(&Hexagon::HvxVRRegClass);
|
2015-11-27 02:38:27 +08:00
|
|
|
RegInfo.addLiveIn(VA.getLocReg(), VReg);
|
|
|
|
InVals.push_back(DAG.getCopyFromReg(Chain, dl, VReg, RegVT));
|
2017-10-19 02:07:07 +08:00
|
|
|
} else if (Subtarget.useHVX128BOps() &&
|
2017-12-15 02:35:24 +08:00
|
|
|
((RegVT == MVT::v32i32 ||
|
2017-10-19 01:45:22 +08:00
|
|
|
RegVT == MVT::v64i16 || RegVT == MVT::v128i8))) {
|
2015-11-27 02:38:27 +08:00
|
|
|
unsigned VReg =
|
2017-09-15 23:46:05 +08:00
|
|
|
RegInfo.createVirtualRegister(&Hexagon::HvxVRRegClass);
|
2015-11-27 02:38:27 +08:00
|
|
|
RegInfo.addLiveIn(VA.getLocReg(), VReg);
|
|
|
|
InVals.push_back(DAG.getCopyFromReg(Chain, dl, VReg, RegVT));
|
|
|
|
|
|
|
|
// Double Vector
|
2017-12-15 02:35:24 +08:00
|
|
|
} else if ((RegVT == MVT::v32i32 ||
|
2015-11-27 02:38:27 +08:00
|
|
|
RegVT == MVT::v64i16 || RegVT == MVT::v128i8)) {
|
|
|
|
unsigned VReg =
|
2017-09-15 23:46:05 +08:00
|
|
|
RegInfo.createVirtualRegister(&Hexagon::HvxWRRegClass);
|
2015-11-27 02:38:27 +08:00
|
|
|
RegInfo.addLiveIn(VA.getLocReg(), VReg);
|
|
|
|
InVals.push_back(DAG.getCopyFromReg(Chain, dl, VReg, RegVT));
|
2017-10-19 02:07:07 +08:00
|
|
|
} else if (Subtarget.useHVX128BOps() &&
|
2017-12-15 02:35:24 +08:00
|
|
|
((RegVT == MVT::v64i32 ||
|
2017-10-19 01:45:22 +08:00
|
|
|
RegVT == MVT::v128i16 || RegVT == MVT::v256i8))) {
|
2015-11-27 02:38:27 +08:00
|
|
|
unsigned VReg =
|
2017-09-15 23:46:05 +08:00
|
|
|
RegInfo.createVirtualRegister(&Hexagon::HvxWRRegClass);
|
2015-11-27 02:38:27 +08:00
|
|
|
RegInfo.addLiveIn(VA.getLocReg(), VReg);
|
|
|
|
InVals.push_back(DAG.getCopyFromReg(Chain, dl, VReg, RegVT));
|
|
|
|
} else if (RegVT == MVT::v512i1 || RegVT == MVT::v1024i1) {
|
|
|
|
assert(0 && "need to support VecPred regs");
|
|
|
|
unsigned VReg =
|
2017-09-15 23:46:05 +08:00
|
|
|
RegInfo.createVirtualRegister(&Hexagon::HvxQRRegClass);
|
2015-11-27 02:38:27 +08:00
|
|
|
RegInfo.addLiveIn(VA.getLocReg(), VReg);
|
|
|
|
InVals.push_back(DAG.getCopyFromReg(Chain, dl, VReg, RegVT));
|
2011-12-13 05:14:40 +08:00
|
|
|
} else {
|
|
|
|
assert (0);
|
|
|
|
}
|
|
|
|
} else if (VA.isRegLoc() && Flags.isByVal() && Flags.getByValSize() <= 8) {
|
|
|
|
assert (0 && "ByValSize must be bigger than 8 bytes");
|
|
|
|
} else {
|
|
|
|
// Sanity check.
|
|
|
|
assert(VA.isMemLoc());
|
|
|
|
|
|
|
|
if (Flags.isByVal()) {
|
|
|
|
// If it's a byval parameter, then we need to compute the
|
|
|
|
// "real" size, not the size of the pointer.
|
|
|
|
ObjSize = Flags.getByValSize();
|
|
|
|
} else {
|
|
|
|
ObjSize = VA.getLocVT().getStoreSizeInBits() >> 3;
|
|
|
|
}
|
|
|
|
|
|
|
|
StackLocation = HEXAGON_LRFP_SIZE + VA.getLocMemOffset();
|
|
|
|
// Create the frame index object for this incoming parameter...
|
2016-07-29 02:40:00 +08:00
|
|
|
FI = MFI.CreateFixedObject(ObjSize, StackLocation, true);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Create the SelectionDAG nodes cordl, responding to a load
|
|
|
|
// from this parameter.
|
|
|
|
SDValue FIN = DAG.getFrameIndex(FI, MVT::i32);
|
|
|
|
|
|
|
|
if (Flags.isByVal()) {
|
|
|
|
// If it's a pass-by-value aggregate, then do not dereference the stack
|
|
|
|
// location. Instead, we should generate a reference to the stack
|
|
|
|
// location.
|
|
|
|
InVals.push_back(FIN);
|
|
|
|
} else {
|
[SelectionDAG] Get rid of bool parameters in SelectionDAG::getLoad, getStore, and friends.
Summary:
Instead, we take a single flags arg (a bitset).
Also add a default 0 alignment, and change the order of arguments so the
alignment comes before the flags.
This greatly simplifies many callsites, and fixes a bug in
AMDGPUISelLowering, wherein the order of the args to getLoad was
inverted. It also greatly simplifies the process of adding another flag
to getLoad.
Reviewers: chandlerc, tstellarAMD
Subscribers: jholewinski, arsenm, jyknight, dsanders, nemanjai, llvm-commits
Differential Revision: http://reviews.llvm.org/D22249
llvm-svn: 275592
2016-07-16 02:27:10 +08:00
|
|
|
InVals.push_back(
|
2017-04-13 23:00:18 +08:00
|
|
|
DAG.getLoad(VA.getValVT(), dl, Chain, FIN, MachinePointerInfo()));
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!MemOps.empty())
|
2014-04-27 02:35:24 +08:00
|
|
|
Chain = DAG.getNode(ISD::TokenFactor, dl, MVT::Other, MemOps);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
if (isVarArg) {
|
|
|
|
// This will point to the next argument passed via stack.
|
2016-07-29 02:40:00 +08:00
|
|
|
int FrameIndex = MFI.CreateFixedObject(Hexagon_PointerSize,
|
|
|
|
HEXAGON_LRFP_SIZE +
|
|
|
|
CCInfo.getNextStackOffset(),
|
|
|
|
true);
|
2015-04-23 00:43:53 +08:00
|
|
|
FuncInfo.setVarArgsFrameIndex(FrameIndex);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return Chain;
|
|
|
|
}
|
|
|
|
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerVASTART(SDValue Op, SelectionDAG &DAG) const {
|
|
|
|
// VASTART stores the address of the VarArgsFrameIndex slot into the
|
|
|
|
// memory location argument.
|
|
|
|
MachineFunction &MF = DAG.getMachineFunction();
|
|
|
|
HexagonMachineFunctionInfo *QFI = MF.getInfo<HexagonMachineFunctionInfo>();
|
|
|
|
SDValue Addr = DAG.getFrameIndex(QFI->getVarArgsFrameIndex(), MVT::i32);
|
|
|
|
const Value *SV = cast<SrcValueSDNode>(Op.getOperand(2))->getValue();
|
[SelectionDAG] Get rid of bool parameters in SelectionDAG::getLoad, getStore, and friends.
Summary:
Instead, we take a single flags arg (a bitset).
Also add a default 0 alignment, and change the order of arguments so the
alignment comes before the flags.
This greatly simplifies many callsites, and fixes a bug in
AMDGPUISelLowering, wherein the order of the args to getLoad was
inverted. It also greatly simplifies the process of adding another flag
to getLoad.
Reviewers: chandlerc, tstellarAMD
Subscribers: jholewinski, arsenm, jyknight, dsanders, nemanjai, llvm-commits
Differential Revision: http://reviews.llvm.org/D22249
llvm-svn: 275592
2016-07-16 02:27:10 +08:00
|
|
|
return DAG.getStore(Op.getOperand(0), SDLoc(Op), Addr, Op.getOperand(1),
|
|
|
|
MachinePointerInfo(SV));
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2015-03-20 00:33:08 +08:00
|
|
|
static bool isSExtFree(SDValue N) {
|
|
|
|
// A sign-extend of a truncate of a sign-extend is free.
|
|
|
|
if (N.getOpcode() == ISD::TRUNCATE &&
|
|
|
|
N.getOperand(0).getOpcode() == ISD::AssertSext)
|
|
|
|
return true;
|
|
|
|
// We have sign-extended loads.
|
|
|
|
if (N.getOpcode() == ISD::LOAD)
|
|
|
|
return true;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
SDValue HexagonTargetLowering::LowerSETCC(SDValue Op, SelectionDAG &DAG) const {
|
|
|
|
SDLoc dl(Op);
|
|
|
|
|
|
|
|
SDValue LHS = Op.getOperand(0);
|
|
|
|
SDValue RHS = Op.getOperand(1);
|
2017-12-15 05:28:48 +08:00
|
|
|
if (Subtarget.useHVXOps() && Subtarget.isHVXVectorType(ty(LHS)))
|
|
|
|
return LowerHvxSetCC(Op, DAG);
|
|
|
|
|
2015-03-20 00:33:08 +08:00
|
|
|
SDValue Cmp = Op.getOperand(2);
|
|
|
|
ISD::CondCode CC = cast<CondCodeSDNode>(Cmp)->get();
|
|
|
|
|
|
|
|
EVT VT = Op.getValueType();
|
|
|
|
EVT LHSVT = LHS.getValueType();
|
|
|
|
EVT RHSVT = RHS.getValueType();
|
|
|
|
|
|
|
|
if (LHSVT == MVT::v2i16) {
|
|
|
|
assert(ISD::isSignedIntSetCC(CC) || ISD::isUnsignedIntSetCC(CC));
|
|
|
|
unsigned ExtOpc = ISD::isSignedIntSetCC(CC) ? ISD::SIGN_EXTEND
|
|
|
|
: ISD::ZERO_EXTEND;
|
|
|
|
SDValue LX = DAG.getNode(ExtOpc, dl, MVT::v2i32, LHS);
|
|
|
|
SDValue RX = DAG.getNode(ExtOpc, dl, MVT::v2i32, RHS);
|
|
|
|
SDValue SC = DAG.getNode(ISD::SETCC, dl, MVT::v2i1, LX, RX, Cmp);
|
|
|
|
return SC;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Treat all other vector types as legal.
|
|
|
|
if (VT.isVector())
|
|
|
|
return Op;
|
|
|
|
|
|
|
|
// Equals and not equals should use sign-extend, not zero-extend, since
|
|
|
|
// we can represent small negative values in the compare instructions.
|
|
|
|
// The LLVM default is to use zero-extend arbitrarily in these cases.
|
|
|
|
if ((CC == ISD::SETEQ || CC == ISD::SETNE) &&
|
|
|
|
(RHSVT == MVT::i8 || RHSVT == MVT::i16) &&
|
|
|
|
(LHSVT == MVT::i8 || LHSVT == MVT::i16)) {
|
|
|
|
ConstantSDNode *C = dyn_cast<ConstantSDNode>(RHS);
|
|
|
|
if (C && C->getAPIntValue().isNegative()) {
|
|
|
|
LHS = DAG.getNode(ISD::SIGN_EXTEND, dl, MVT::i32, LHS);
|
|
|
|
RHS = DAG.getNode(ISD::SIGN_EXTEND, dl, MVT::i32, RHS);
|
|
|
|
return DAG.getNode(ISD::SETCC, dl, Op.getValueType(),
|
|
|
|
LHS, RHS, Op.getOperand(2));
|
|
|
|
}
|
|
|
|
if (isSExtFree(LHS) || isSExtFree(RHS)) {
|
|
|
|
LHS = DAG.getNode(ISD::SIGN_EXTEND, dl, MVT::i32, LHS);
|
|
|
|
RHS = DAG.getNode(ISD::SIGN_EXTEND, dl, MVT::i32, RHS);
|
|
|
|
return DAG.getNode(ISD::SETCC, dl, Op.getValueType(),
|
|
|
|
LHS, RHS, Op.getOperand(2));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return SDValue();
|
|
|
|
}
|
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerVSELECT(SDValue Op, SelectionDAG &DAG) const {
|
2015-03-20 00:33:08 +08:00
|
|
|
SDValue PredOp = Op.getOperand(0);
|
|
|
|
SDValue Op1 = Op.getOperand(1), Op2 = Op.getOperand(2);
|
|
|
|
EVT OpVT = Op1.getValueType();
|
|
|
|
SDLoc DL(Op);
|
|
|
|
|
|
|
|
if (OpVT == MVT::v2i16) {
|
|
|
|
SDValue X1 = DAG.getNode(ISD::ZERO_EXTEND, DL, MVT::v2i32, Op1);
|
|
|
|
SDValue X2 = DAG.getNode(ISD::ZERO_EXTEND, DL, MVT::v2i32, Op2);
|
|
|
|
SDValue SL = DAG.getNode(ISD::VSELECT, DL, MVT::v2i32, PredOp, X1, X2);
|
|
|
|
SDValue TR = DAG.getNode(ISD::TRUNCATE, DL, MVT::v2i16, SL);
|
|
|
|
return TR;
|
|
|
|
}
|
|
|
|
|
|
|
|
return SDValue();
|
|
|
|
}
|
|
|
|
|
2017-08-01 21:12:53 +08:00
|
|
|
static Constant *convert_i1_to_i8(const Constant *ConstVal) {
|
|
|
|
SmallVector<Constant *, 128> NewConst;
|
|
|
|
const ConstantVector *CV = dyn_cast<ConstantVector>(ConstVal);
|
|
|
|
if (!CV)
|
|
|
|
return nullptr;
|
|
|
|
|
|
|
|
LLVMContext &Ctx = ConstVal->getContext();
|
|
|
|
IRBuilder<> IRB(Ctx);
|
|
|
|
unsigned NumVectorElements = CV->getNumOperands();
|
|
|
|
assert(isPowerOf2_32(NumVectorElements) &&
|
|
|
|
"conversion only supported for pow2 VectorSize!");
|
|
|
|
|
|
|
|
for (unsigned i = 0; i < NumVectorElements / 8; ++i) {
|
|
|
|
uint8_t x = 0;
|
|
|
|
for (unsigned j = 0; j < 8; ++j) {
|
|
|
|
uint8_t y = CV->getOperand(i * 8 + j)->getUniqueInteger().getZExtValue();
|
|
|
|
x |= y << (7 - j);
|
|
|
|
}
|
|
|
|
assert((x == 0 || x == 255) && "Either all 0's or all 1's expected!");
|
|
|
|
NewConst.push_back(IRB.getInt8(x));
|
|
|
|
}
|
|
|
|
return ConstantVector::get(NewConst);
|
|
|
|
}
|
|
|
|
|
2012-05-11 04:20:25 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerConstantPool(SDValue Op, SelectionDAG &DAG) const {
|
|
|
|
EVT ValTy = Op.getValueType();
|
2015-12-19 04:19:30 +08:00
|
|
|
ConstantPoolSDNode *CPN = cast<ConstantPoolSDNode>(Op);
|
2017-08-01 21:12:53 +08:00
|
|
|
Constant *CVal = nullptr;
|
|
|
|
bool isVTi1Type = false;
|
|
|
|
if (const Constant *ConstVal = dyn_cast<Constant>(CPN->getConstVal())) {
|
|
|
|
Type *CValTy = ConstVal->getType();
|
|
|
|
if (CValTy->isVectorTy() &&
|
|
|
|
CValTy->getVectorElementType()->isIntegerTy(1)) {
|
|
|
|
CVal = convert_i1_to_i8(ConstVal);
|
|
|
|
isVTi1Type = (CVal != nullptr);
|
|
|
|
}
|
|
|
|
}
|
2015-12-19 04:19:30 +08:00
|
|
|
unsigned Align = CPN->getAlignment();
|
2016-06-27 06:24:01 +08:00
|
|
|
bool IsPositionIndependent = isPositionIndependent();
|
|
|
|
unsigned char TF = IsPositionIndependent ? HexagonII::MO_PCREL : 0;
|
2015-12-19 04:19:30 +08:00
|
|
|
|
2016-08-14 07:41:11 +08:00
|
|
|
unsigned Offset = 0;
|
2015-12-19 04:19:30 +08:00
|
|
|
SDValue T;
|
|
|
|
if (CPN->isMachineConstantPoolEntry())
|
2016-08-14 07:41:11 +08:00
|
|
|
T = DAG.getTargetConstantPool(CPN->getMachineCPVal(), ValTy, Align, Offset,
|
|
|
|
TF);
|
2017-08-01 21:12:53 +08:00
|
|
|
else if (isVTi1Type)
|
|
|
|
T = DAG.getTargetConstantPool(CVal, ValTy, Align, Offset, TF);
|
2012-05-11 04:20:25 +08:00
|
|
|
else
|
2016-08-14 07:41:11 +08:00
|
|
|
T = DAG.getTargetConstantPool(CPN->getConstVal(), ValTy, Align, Offset,
|
|
|
|
TF);
|
|
|
|
|
|
|
|
assert(cast<ConstantPoolSDNode>(T)->getTargetFlags() == TF &&
|
|
|
|
"Inconsistent target flag encountered");
|
|
|
|
|
2016-06-27 06:24:01 +08:00
|
|
|
if (IsPositionIndependent)
|
2015-12-19 04:19:30 +08:00
|
|
|
return DAG.getNode(HexagonISD::AT_PCREL, SDLoc(Op), ValTy, T);
|
|
|
|
return DAG.getNode(HexagonISD::CP, SDLoc(Op), ValTy, T);
|
|
|
|
}
|
|
|
|
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerJumpTable(SDValue Op, SelectionDAG &DAG) const {
|
|
|
|
EVT VT = Op.getValueType();
|
|
|
|
int Idx = cast<JumpTableSDNode>(Op)->getIndex();
|
2016-06-27 06:24:01 +08:00
|
|
|
if (isPositionIndependent()) {
|
2015-12-19 04:19:30 +08:00
|
|
|
SDValue T = DAG.getTargetJumpTable(Idx, VT, HexagonII::MO_PCREL);
|
|
|
|
return DAG.getNode(HexagonISD::AT_PCREL, SDLoc(Op), VT, T);
|
|
|
|
}
|
|
|
|
|
|
|
|
SDValue T = DAG.getTargetJumpTable(Idx, VT);
|
|
|
|
return DAG.getNode(HexagonISD::JT, SDLoc(Op), VT, T);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerRETURNADDR(SDValue Op, SelectionDAG &DAG) const {
|
2015-04-23 05:17:00 +08:00
|
|
|
const HexagonRegisterInfo &HRI = *Subtarget.getRegisterInfo();
|
2011-12-13 05:14:40 +08:00
|
|
|
MachineFunction &MF = DAG.getMachineFunction();
|
2016-07-29 02:40:00 +08:00
|
|
|
MachineFrameInfo &MFI = MF.getFrameInfo();
|
2015-04-23 05:17:00 +08:00
|
|
|
MFI.setReturnAddressIsTaken(true);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2014-01-06 08:43:20 +08:00
|
|
|
if (verifyReturnAddressArgumentIsConstant(Op, DAG))
|
2014-01-05 09:47:20 +08:00
|
|
|
return SDValue();
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
EVT VT = Op.getValueType();
|
2013-05-25 10:42:55 +08:00
|
|
|
SDLoc dl(Op);
|
2011-12-13 05:14:40 +08:00
|
|
|
unsigned Depth = cast<ConstantSDNode>(Op.getOperand(0))->getZExtValue();
|
|
|
|
if (Depth) {
|
|
|
|
SDValue FrameAddr = LowerFRAMEADDR(Op, DAG);
|
2015-04-28 22:05:47 +08:00
|
|
|
SDValue Offset = DAG.getConstant(4, dl, MVT::i32);
|
2011-12-13 05:14:40 +08:00
|
|
|
return DAG.getLoad(VT, dl, DAG.getEntryNode(),
|
|
|
|
DAG.getNode(ISD::ADD, dl, VT, FrameAddr, Offset),
|
[SelectionDAG] Get rid of bool parameters in SelectionDAG::getLoad, getStore, and friends.
Summary:
Instead, we take a single flags arg (a bitset).
Also add a default 0 alignment, and change the order of arguments so the
alignment comes before the flags.
This greatly simplifies many callsites, and fixes a bug in
AMDGPUISelLowering, wherein the order of the args to getLoad was
inverted. It also greatly simplifies the process of adding another flag
to getLoad.
Reviewers: chandlerc, tstellarAMD
Subscribers: jholewinski, arsenm, jyknight, dsanders, nemanjai, llvm-commits
Differential Revision: http://reviews.llvm.org/D22249
llvm-svn: 275592
2016-07-16 02:27:10 +08:00
|
|
|
MachinePointerInfo());
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Return LR, which contains the return address. Mark it an implicit live-in.
|
2015-04-23 05:17:00 +08:00
|
|
|
unsigned Reg = MF.addLiveIn(HRI.getRARegister(), getRegClassFor(MVT::i32));
|
2011-12-13 05:14:40 +08:00
|
|
|
return DAG.getCopyFromReg(DAG.getEntryNode(), dl, Reg, VT);
|
|
|
|
}
|
|
|
|
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerFRAMEADDR(SDValue Op, SelectionDAG &DAG) const {
|
2015-04-23 05:17:00 +08:00
|
|
|
const HexagonRegisterInfo &HRI = *Subtarget.getRegisterInfo();
|
2016-07-29 02:40:00 +08:00
|
|
|
MachineFrameInfo &MFI = DAG.getMachineFunction().getFrameInfo();
|
2015-04-23 05:17:00 +08:00
|
|
|
MFI.setFrameAddressIsTaken(true);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
EVT VT = Op.getValueType();
|
2013-05-25 10:42:55 +08:00
|
|
|
SDLoc dl(Op);
|
2011-12-13 05:14:40 +08:00
|
|
|
unsigned Depth = cast<ConstantSDNode>(Op.getOperand(0))->getZExtValue();
|
|
|
|
SDValue FrameAddr = DAG.getCopyFromReg(DAG.getEntryNode(), dl,
|
2015-04-23 05:17:00 +08:00
|
|
|
HRI.getFrameRegister(), VT);
|
2011-12-13 05:14:40 +08:00
|
|
|
while (Depth--)
|
|
|
|
FrameAddr = DAG.getLoad(VT, dl, DAG.getEntryNode(), FrameAddr,
|
[SelectionDAG] Get rid of bool parameters in SelectionDAG::getLoad, getStore, and friends.
Summary:
Instead, we take a single flags arg (a bitset).
Also add a default 0 alignment, and change the order of arguments so the
alignment comes before the flags.
This greatly simplifies many callsites, and fixes a bug in
AMDGPUISelLowering, wherein the order of the args to getLoad was
inverted. It also greatly simplifies the process of adding another flag
to getLoad.
Reviewers: chandlerc, tstellarAMD
Subscribers: jholewinski, arsenm, jyknight, dsanders, nemanjai, llvm-commits
Differential Revision: http://reviews.llvm.org/D22249
llvm-svn: 275592
2016-07-16 02:27:10 +08:00
|
|
|
MachinePointerInfo());
|
2011-12-13 05:14:40 +08:00
|
|
|
return FrameAddr;
|
|
|
|
}
|
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerATOMIC_FENCE(SDValue Op, SelectionDAG& DAG) const {
|
2013-05-25 10:42:55 +08:00
|
|
|
SDLoc dl(Op);
|
2011-12-13 05:14:40 +08:00
|
|
|
return DAG.getNode(HexagonISD::BARRIER, dl, MVT::Other, Op.getOperand(0));
|
|
|
|
}
|
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerGLOBALADDRESS(SDValue Op, SelectionDAG &DAG) const {
|
2013-05-25 10:42:55 +08:00
|
|
|
SDLoc dl(Op);
|
2015-12-19 04:19:30 +08:00
|
|
|
auto *GAN = cast<GlobalAddressSDNode>(Op);
|
2015-07-09 10:09:04 +08:00
|
|
|
auto PtrVT = getPointerTy(DAG.getDataLayout());
|
2015-12-19 04:19:30 +08:00
|
|
|
auto *GV = GAN->getGlobal();
|
|
|
|
int64_t Offset = GAN->getOffset();
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
auto &HLOF = *HTM.getObjFileLowering();
|
|
|
|
Reloc::Model RM = HTM.getRelocationModel();
|
|
|
|
|
|
|
|
if (RM == Reloc::Static) {
|
|
|
|
SDValue GA = DAG.getTargetGlobalAddress(GV, dl, PtrVT, Offset);
|
2016-10-25 03:23:39 +08:00
|
|
|
const GlobalObject *GO = GV->getBaseObject();
|
|
|
|
if (GO && HLOF.isGlobalInSmallSection(GO, HTM))
|
2015-12-19 04:19:30 +08:00
|
|
|
return DAG.getNode(HexagonISD::CONST32_GP, dl, PtrVT, GA);
|
|
|
|
return DAG.getNode(HexagonISD::CONST32, dl, PtrVT, GA);
|
|
|
|
}
|
|
|
|
|
2016-06-28 07:15:57 +08:00
|
|
|
bool UsePCRel = getTargetMachine().shouldAssumeDSOLocal(*GV->getParent(), GV);
|
2015-12-19 04:19:30 +08:00
|
|
|
if (UsePCRel) {
|
|
|
|
SDValue GA = DAG.getTargetGlobalAddress(GV, dl, PtrVT, Offset,
|
|
|
|
HexagonII::MO_PCREL);
|
|
|
|
return DAG.getNode(HexagonISD::AT_PCREL, dl, PtrVT, GA);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
// Use GOT index.
|
|
|
|
SDValue GOT = DAG.getGLOBAL_OFFSET_TABLE(PtrVT);
|
|
|
|
SDValue GA = DAG.getTargetGlobalAddress(GV, dl, PtrVT, 0, HexagonII::MO_GOT);
|
|
|
|
SDValue Off = DAG.getConstant(Offset, dl, MVT::i32);
|
|
|
|
return DAG.getNode(HexagonISD::AT_GOT, dl, PtrVT, GOT, GA, Off);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2015-03-20 00:33:08 +08:00
|
|
|
// Specifies that for loads and stores VT can be promoted to PromotedLdStVT.
|
2015-12-19 04:19:30 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerBlockAddress(SDValue Op, SelectionDAG &DAG) const {
|
|
|
|
const BlockAddress *BA = cast<BlockAddressSDNode>(Op)->getBlockAddress();
|
|
|
|
SDLoc dl(Op);
|
|
|
|
EVT PtrVT = getPointerTy(DAG.getDataLayout());
|
2015-03-20 00:33:08 +08:00
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
Reloc::Model RM = HTM.getRelocationModel();
|
|
|
|
if (RM == Reloc::Static) {
|
2016-02-18 23:42:57 +08:00
|
|
|
SDValue A = DAG.getTargetBlockAddress(BA, PtrVT);
|
2015-12-19 04:19:30 +08:00
|
|
|
return DAG.getNode(HexagonISD::CONST32_GP, dl, PtrVT, A);
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
2015-12-19 04:19:30 +08:00
|
|
|
|
|
|
|
SDValue A = DAG.getTargetBlockAddress(BA, PtrVT, 0, HexagonII::MO_PCREL);
|
|
|
|
return DAG.getNode(HexagonISD::AT_PCREL, dl, PtrVT, A);
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
|
|
|
|
2013-03-08 03:10:28 +08:00
|
|
|
SDValue
|
2015-12-19 04:19:30 +08:00
|
|
|
HexagonTargetLowering::LowerGLOBAL_OFFSET_TABLE(SDValue Op, SelectionDAG &DAG)
|
|
|
|
const {
|
|
|
|
EVT PtrVT = getPointerTy(DAG.getDataLayout());
|
|
|
|
SDValue GOTSym = DAG.getTargetExternalSymbol(HEXAGON_GOT_SYM_NAME, PtrVT,
|
|
|
|
HexagonII::MO_PCREL);
|
|
|
|
return DAG.getNode(HexagonISD::AT_PCREL, SDLoc(Op), PtrVT, GOTSym);
|
2013-03-08 03:10:28 +08:00
|
|
|
}
|
|
|
|
|
2016-02-18 23:42:57 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::GetDynamicTLSAddr(SelectionDAG &DAG, SDValue Chain,
|
2017-02-18 06:14:51 +08:00
|
|
|
GlobalAddressSDNode *GA, SDValue Glue, EVT PtrVT, unsigned ReturnReg,
|
2016-02-18 23:42:57 +08:00
|
|
|
unsigned char OperandFlags) const {
|
2017-02-18 06:14:51 +08:00
|
|
|
MachineFunction &MF = DAG.getMachineFunction();
|
|
|
|
MachineFrameInfo &MFI = MF.getFrameInfo();
|
2016-02-18 23:42:57 +08:00
|
|
|
SDVTList NodeTys = DAG.getVTList(MVT::Other, MVT::Glue);
|
|
|
|
SDLoc dl(GA);
|
|
|
|
SDValue TGA = DAG.getTargetGlobalAddress(GA->getGlobal(), dl,
|
|
|
|
GA->getValueType(0),
|
|
|
|
GA->getOffset(),
|
|
|
|
OperandFlags);
|
|
|
|
// Create Operands for the call.The Operands should have the following:
|
|
|
|
// 1. Chain SDValue
|
|
|
|
// 2. Callee which in this case is the Global address value.
|
|
|
|
// 3. Registers live into the call.In this case its R0, as we
|
|
|
|
// have just one argument to be passed.
|
2017-02-18 06:14:51 +08:00
|
|
|
// 4. Glue.
|
2016-02-18 23:42:57 +08:00
|
|
|
// Note: The order is important.
|
|
|
|
|
2017-02-18 06:14:51 +08:00
|
|
|
const auto &HRI = *Subtarget.getRegisterInfo();
|
|
|
|
const uint32_t *Mask = HRI.getCallPreservedMask(MF, CallingConv::C);
|
|
|
|
assert(Mask && "Missing call preserved mask for calling convention");
|
|
|
|
SDValue Ops[] = { Chain, TGA, DAG.getRegister(Hexagon::R0, PtrVT),
|
|
|
|
DAG.getRegisterMask(Mask), Glue };
|
|
|
|
Chain = DAG.getNode(HexagonISD::CALL, dl, NodeTys, Ops);
|
2016-02-18 23:42:57 +08:00
|
|
|
|
|
|
|
// Inform MFI that function has calls.
|
2016-07-29 02:40:00 +08:00
|
|
|
MFI.setAdjustsStack(true);
|
2016-02-18 23:42:57 +08:00
|
|
|
|
2017-02-18 06:14:51 +08:00
|
|
|
Glue = Chain.getValue(1);
|
|
|
|
return DAG.getCopyFromReg(Chain, dl, ReturnReg, PtrVT, Glue);
|
2016-02-18 23:42:57 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Lower using the intial executable model for TLS addresses
|
|
|
|
//
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerToTLSInitialExecModel(GlobalAddressSDNode *GA,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
SDLoc dl(GA);
|
|
|
|
int64_t Offset = GA->getOffset();
|
|
|
|
auto PtrVT = getPointerTy(DAG.getDataLayout());
|
|
|
|
|
|
|
|
// Get the thread pointer.
|
|
|
|
SDValue TP = DAG.getCopyFromReg(DAG.getEntryNode(), dl, Hexagon::UGP, PtrVT);
|
|
|
|
|
2016-06-27 06:24:01 +08:00
|
|
|
bool IsPositionIndependent = isPositionIndependent();
|
|
|
|
unsigned char TF =
|
|
|
|
IsPositionIndependent ? HexagonII::MO_IEGOT : HexagonII::MO_IE;
|
2016-02-18 23:42:57 +08:00
|
|
|
|
|
|
|
// First generate the TLS symbol address
|
|
|
|
SDValue TGA = DAG.getTargetGlobalAddress(GA->getGlobal(), dl, PtrVT,
|
|
|
|
Offset, TF);
|
|
|
|
|
|
|
|
SDValue Sym = DAG.getNode(HexagonISD::CONST32, dl, PtrVT, TGA);
|
|
|
|
|
2016-06-27 06:24:01 +08:00
|
|
|
if (IsPositionIndependent) {
|
2016-02-18 23:42:57 +08:00
|
|
|
// Generate the GOT pointer in case of position independent code
|
|
|
|
SDValue GOT = LowerGLOBAL_OFFSET_TABLE(Sym, DAG);
|
|
|
|
|
|
|
|
// Add the TLS Symbol address to GOT pointer.This gives
|
|
|
|
// GOT relative relocation for the symbol.
|
|
|
|
Sym = DAG.getNode(ISD::ADD, dl, PtrVT, GOT, Sym);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load the offset value for TLS symbol.This offset is relative to
|
|
|
|
// thread pointer.
|
[SelectionDAG] Get rid of bool parameters in SelectionDAG::getLoad, getStore, and friends.
Summary:
Instead, we take a single flags arg (a bitset).
Also add a default 0 alignment, and change the order of arguments so the
alignment comes before the flags.
This greatly simplifies many callsites, and fixes a bug in
AMDGPUISelLowering, wherein the order of the args to getLoad was
inverted. It also greatly simplifies the process of adding another flag
to getLoad.
Reviewers: chandlerc, tstellarAMD
Subscribers: jholewinski, arsenm, jyknight, dsanders, nemanjai, llvm-commits
Differential Revision: http://reviews.llvm.org/D22249
llvm-svn: 275592
2016-07-16 02:27:10 +08:00
|
|
|
SDValue LoadOffset =
|
|
|
|
DAG.getLoad(PtrVT, dl, DAG.getEntryNode(), Sym, MachinePointerInfo());
|
2016-02-18 23:42:57 +08:00
|
|
|
|
|
|
|
// Address of the thread local variable is the add of thread
|
|
|
|
// pointer and the offset of the variable.
|
|
|
|
return DAG.getNode(ISD::ADD, dl, PtrVT, TP, LoadOffset);
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Lower using the local executable model for TLS addresses
|
|
|
|
//
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerToTLSLocalExecModel(GlobalAddressSDNode *GA,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
SDLoc dl(GA);
|
|
|
|
int64_t Offset = GA->getOffset();
|
|
|
|
auto PtrVT = getPointerTy(DAG.getDataLayout());
|
|
|
|
|
|
|
|
// Get the thread pointer.
|
|
|
|
SDValue TP = DAG.getCopyFromReg(DAG.getEntryNode(), dl, Hexagon::UGP, PtrVT);
|
|
|
|
// Generate the TLS symbol address
|
|
|
|
SDValue TGA = DAG.getTargetGlobalAddress(GA->getGlobal(), dl, PtrVT, Offset,
|
|
|
|
HexagonII::MO_TPREL);
|
|
|
|
SDValue Sym = DAG.getNode(HexagonISD::CONST32, dl, PtrVT, TGA);
|
|
|
|
|
|
|
|
// Address of the thread local variable is the add of thread
|
|
|
|
// pointer and the offset of the variable.
|
|
|
|
return DAG.getNode(ISD::ADD, dl, PtrVT, TP, Sym);
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Lower using the general dynamic model for TLS addresses
|
|
|
|
//
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerToTLSGeneralDynamicModel(GlobalAddressSDNode *GA,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
SDLoc dl(GA);
|
|
|
|
int64_t Offset = GA->getOffset();
|
|
|
|
auto PtrVT = getPointerTy(DAG.getDataLayout());
|
|
|
|
|
|
|
|
// First generate the TLS symbol address
|
|
|
|
SDValue TGA = DAG.getTargetGlobalAddress(GA->getGlobal(), dl, PtrVT, Offset,
|
|
|
|
HexagonII::MO_GDGOT);
|
|
|
|
|
|
|
|
// Then, generate the GOT pointer
|
|
|
|
SDValue GOT = LowerGLOBAL_OFFSET_TABLE(TGA, DAG);
|
|
|
|
|
|
|
|
// Add the TLS symbol and the GOT pointer
|
|
|
|
SDValue Sym = DAG.getNode(HexagonISD::CONST32, dl, PtrVT, TGA);
|
|
|
|
SDValue Chain = DAG.getNode(ISD::ADD, dl, PtrVT, GOT, Sym);
|
|
|
|
|
|
|
|
// Copy over the argument to R0
|
|
|
|
SDValue InFlag;
|
|
|
|
Chain = DAG.getCopyToReg(DAG.getEntryNode(), dl, Hexagon::R0, Chain, InFlag);
|
|
|
|
InFlag = Chain.getValue(1);
|
|
|
|
|
2017-05-03 02:15:33 +08:00
|
|
|
unsigned Flags =
|
|
|
|
static_cast<const HexagonSubtarget &>(DAG.getSubtarget()).useLongCalls()
|
|
|
|
? HexagonII::MO_GDPLT | HexagonII::HMOTF_ConstExtended
|
|
|
|
: HexagonII::MO_GDPLT;
|
|
|
|
|
2017-02-18 06:14:51 +08:00
|
|
|
return GetDynamicTLSAddr(DAG, Chain, GA, InFlag, PtrVT,
|
2017-05-03 02:15:33 +08:00
|
|
|
Hexagon::R0, Flags);
|
2016-02-18 23:42:57 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Lower TLS addresses.
|
|
|
|
//
|
|
|
|
// For now for dynamic models, we only support the general dynamic model.
|
|
|
|
//
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerGlobalTLSAddress(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
GlobalAddressSDNode *GA = cast<GlobalAddressSDNode>(Op);
|
|
|
|
|
|
|
|
switch (HTM.getTLSModel(GA->getGlobal())) {
|
|
|
|
case TLSModel::GeneralDynamic:
|
|
|
|
case TLSModel::LocalDynamic:
|
|
|
|
return LowerToTLSGeneralDynamicModel(GA, DAG);
|
|
|
|
case TLSModel::InitialExec:
|
|
|
|
return LowerToTLSInitialExecModel(GA, DAG);
|
|
|
|
case TLSModel::LocalExec:
|
|
|
|
return LowerToTLSLocalExecModel(GA, DAG);
|
|
|
|
}
|
|
|
|
llvm_unreachable("Bogus TLS model");
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// TargetLowering Implementation
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2015-02-03 06:11:36 +08:00
|
|
|
HexagonTargetLowering::HexagonTargetLowering(const TargetMachine &TM,
|
2015-11-27 02:38:27 +08:00
|
|
|
const HexagonSubtarget &ST)
|
2015-04-23 05:17:00 +08:00
|
|
|
: TargetLowering(TM), HTM(static_cast<const HexagonTargetMachine&>(TM)),
|
2015-11-27 02:38:27 +08:00
|
|
|
Subtarget(ST) {
|
2015-04-23 05:17:00 +08:00
|
|
|
bool IsV4 = !Subtarget.hasV5TOps();
|
|
|
|
auto &HRI = *Subtarget.getRegisterInfo();
|
2015-03-20 00:33:08 +08:00
|
|
|
|
2014-06-27 08:13:49 +08:00
|
|
|
setPrefLoopAlignment(4);
|
2015-04-23 05:17:00 +08:00
|
|
|
setPrefFunctionAlignment(4);
|
|
|
|
setMinFunctionAlignment(2);
|
|
|
|
setStackPointerRegisterToSaveRestore(HRI.getStackRegister());
|
|
|
|
|
2016-06-23 00:07:10 +08:00
|
|
|
setMaxAtomicSizeInBitsSupported(64);
|
|
|
|
setMinCmpXchgSizeInBits(32);
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
if (EnableHexSDNodeSched)
|
|
|
|
setSchedulingPreference(Sched::VLIW);
|
|
|
|
else
|
|
|
|
setSchedulingPreference(Sched::Source);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2014-06-27 08:13:49 +08:00
|
|
|
// Limits for inline expansion of memcpy/memmove
|
2015-04-23 05:17:00 +08:00
|
|
|
MaxStoresPerMemcpy = MaxStoresPerMemcpyCL;
|
|
|
|
MaxStoresPerMemcpyOptSize = MaxStoresPerMemcpyOptSizeCL;
|
|
|
|
MaxStoresPerMemmove = MaxStoresPerMemmoveCL;
|
|
|
|
MaxStoresPerMemmoveOptSize = MaxStoresPerMemmoveOptSizeCL;
|
|
|
|
MaxStoresPerMemset = MaxStoresPerMemsetCL;
|
|
|
|
MaxStoresPerMemsetOptSize = MaxStoresPerMemsetOptSizeCL;
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2014-06-27 08:13:49 +08:00
|
|
|
//
|
2015-04-23 05:17:00 +08:00
|
|
|
// Set up register classes.
|
2014-06-27 08:13:49 +08:00
|
|
|
//
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
addRegisterClass(MVT::i1, &Hexagon::PredRegsRegClass);
|
|
|
|
addRegisterClass(MVT::v2i1, &Hexagon::PredRegsRegClass); // bbbbaaaa
|
|
|
|
addRegisterClass(MVT::v4i1, &Hexagon::PredRegsRegClass); // ddccbbaa
|
|
|
|
addRegisterClass(MVT::v8i1, &Hexagon::PredRegsRegClass); // hgfedcba
|
|
|
|
addRegisterClass(MVT::i32, &Hexagon::IntRegsRegClass);
|
|
|
|
addRegisterClass(MVT::v4i8, &Hexagon::IntRegsRegClass);
|
|
|
|
addRegisterClass(MVT::v2i16, &Hexagon::IntRegsRegClass);
|
|
|
|
addRegisterClass(MVT::i64, &Hexagon::DoubleRegsRegClass);
|
|
|
|
addRegisterClass(MVT::v8i8, &Hexagon::DoubleRegsRegClass);
|
|
|
|
addRegisterClass(MVT::v4i16, &Hexagon::DoubleRegsRegClass);
|
|
|
|
addRegisterClass(MVT::v2i32, &Hexagon::DoubleRegsRegClass);
|
2012-05-11 04:20:25 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
if (Subtarget.hasV5TOps()) {
|
|
|
|
addRegisterClass(MVT::f32, &Hexagon::IntRegsRegClass);
|
|
|
|
addRegisterClass(MVT::f64, &Hexagon::DoubleRegsRegClass);
|
2014-06-27 08:13:49 +08:00
|
|
|
}
|
2012-05-11 04:20:25 +08:00
|
|
|
|
2015-11-26 12:33:11 +08:00
|
|
|
if (Subtarget.hasV60TOps()) {
|
2017-10-19 02:07:07 +08:00
|
|
|
if (Subtarget.useHVX64BOps()) {
|
2017-09-15 23:46:05 +08:00
|
|
|
addRegisterClass(MVT::v64i8, &Hexagon::HvxVRRegClass);
|
|
|
|
addRegisterClass(MVT::v32i16, &Hexagon::HvxVRRegClass);
|
|
|
|
addRegisterClass(MVT::v16i32, &Hexagon::HvxVRRegClass);
|
|
|
|
addRegisterClass(MVT::v128i8, &Hexagon::HvxWRRegClass);
|
|
|
|
addRegisterClass(MVT::v64i16, &Hexagon::HvxWRRegClass);
|
|
|
|
addRegisterClass(MVT::v32i32, &Hexagon::HvxWRRegClass);
|
2017-12-15 05:28:48 +08:00
|
|
|
addRegisterClass(MVT::v16i1, &Hexagon::HvxQRRegClass);
|
|
|
|
addRegisterClass(MVT::v32i1, &Hexagon::HvxQRRegClass);
|
|
|
|
addRegisterClass(MVT::v64i1, &Hexagon::HvxQRRegClass);
|
2017-09-15 23:46:05 +08:00
|
|
|
addRegisterClass(MVT::v512i1, &Hexagon::HvxQRRegClass);
|
2017-10-19 02:07:07 +08:00
|
|
|
} else if (Subtarget.useHVX128BOps()) {
|
2017-09-15 23:46:05 +08:00
|
|
|
addRegisterClass(MVT::v128i8, &Hexagon::HvxVRRegClass);
|
|
|
|
addRegisterClass(MVT::v64i16, &Hexagon::HvxVRRegClass);
|
|
|
|
addRegisterClass(MVT::v32i32, &Hexagon::HvxVRRegClass);
|
|
|
|
addRegisterClass(MVT::v256i8, &Hexagon::HvxWRRegClass);
|
|
|
|
addRegisterClass(MVT::v128i16, &Hexagon::HvxWRRegClass);
|
|
|
|
addRegisterClass(MVT::v64i32, &Hexagon::HvxWRRegClass);
|
2017-12-15 05:28:48 +08:00
|
|
|
addRegisterClass(MVT::v32i1, &Hexagon::HvxQRRegClass);
|
|
|
|
addRegisterClass(MVT::v64i1, &Hexagon::HvxQRRegClass);
|
|
|
|
addRegisterClass(MVT::v128i1, &Hexagon::HvxQRRegClass);
|
2017-09-15 23:46:05 +08:00
|
|
|
addRegisterClass(MVT::v1024i1, &Hexagon::HvxQRRegClass);
|
2015-11-26 12:33:11 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
//
|
|
|
|
// Handling of scalar operations.
|
|
|
|
//
|
|
|
|
// All operations default to "legal", except:
|
|
|
|
// - indexed loads and stores (pre-/post-incremented),
|
|
|
|
// - ANY_EXTEND_VECTOR_INREG, ATOMIC_CMP_SWAP_WITH_SUCCESS, CONCAT_VECTORS,
|
|
|
|
// ConstantFP, DEBUGTRAP, FCEIL, FCOPYSIGN, FEXP, FEXP2, FFLOOR, FGETSIGN,
|
|
|
|
// FLOG, FLOG2, FLOG10, FMAXNUM, FMINNUM, FNEARBYINT, FRINT, FROUND, TRAP,
|
|
|
|
// FTRUNC, PREFETCH, SIGN_EXTEND_VECTOR_INREG, ZERO_EXTEND_VECTOR_INREG,
|
|
|
|
// which default to "expand" for at least one type.
|
|
|
|
|
|
|
|
// Misc operations.
|
|
|
|
setOperationAction(ISD::ConstantFP, MVT::f32, Legal); // Default: expand
|
|
|
|
setOperationAction(ISD::ConstantFP, MVT::f64, Legal); // Default: expand
|
2014-06-27 08:13:49 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::ConstantPool, MVT::i32, Custom);
|
2015-12-19 04:19:30 +08:00
|
|
|
setOperationAction(ISD::JumpTable, MVT::i32, Custom);
|
2014-06-27 08:13:49 +08:00
|
|
|
setOperationAction(ISD::BUILD_PAIR, MVT::i64, Expand);
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::SIGN_EXTEND_INREG, MVT::i1, Expand);
|
|
|
|
setOperationAction(ISD::INLINEASM, MVT::Other, Custom);
|
2016-02-18 21:58:38 +08:00
|
|
|
setOperationAction(ISD::PREFETCH, MVT::Other, Custom);
|
2017-02-23 06:28:47 +08:00
|
|
|
setOperationAction(ISD::READCYCLECOUNTER, MVT::i64, Custom);
|
2016-02-18 21:58:38 +08:00
|
|
|
setOperationAction(ISD::INTRINSIC_VOID, MVT::Other, Custom);
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::EH_RETURN, MVT::Other, Custom);
|
2015-12-19 04:19:30 +08:00
|
|
|
setOperationAction(ISD::GLOBAL_OFFSET_TABLE, MVT::i32, Custom);
|
2016-02-18 23:42:57 +08:00
|
|
|
setOperationAction(ISD::GlobalTLSAddress, MVT::i32, Custom);
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::ATOMIC_FENCE, MVT::Other, Custom);
|
2014-06-27 08:13:49 +08:00
|
|
|
|
|
|
|
// Custom legalize GlobalAddress nodes into CONST32.
|
|
|
|
setOperationAction(ISD::GlobalAddress, MVT::i32, Custom);
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::GlobalAddress, MVT::i8, Custom);
|
|
|
|
setOperationAction(ISD::BlockAddress, MVT::i32, Custom);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-03-20 00:33:08 +08:00
|
|
|
// Hexagon needs to optimize cases with negative constants.
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::SETCC, MVT::i8, Custom);
|
2015-03-20 00:33:08 +08:00
|
|
|
setOperationAction(ISD::SETCC, MVT::i16, Custom);
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
// VASTART needs to be custom lowered to use the VarArgsFrameIndex.
|
|
|
|
setOperationAction(ISD::VASTART, MVT::Other, Custom);
|
|
|
|
setOperationAction(ISD::VAEND, MVT::Other, Expand);
|
|
|
|
setOperationAction(ISD::VAARG, MVT::Other, Expand);
|
|
|
|
|
|
|
|
setOperationAction(ISD::STACKSAVE, MVT::Other, Expand);
|
|
|
|
setOperationAction(ISD::STACKRESTORE, MVT::Other, Expand);
|
|
|
|
setOperationAction(ISD::DYNAMIC_STACKALLOC, MVT::i32, Custom);
|
|
|
|
|
|
|
|
if (EmitJumpTables)
|
2015-12-19 04:19:30 +08:00
|
|
|
setMinimumJumpTableEntries(MinimumJumpTables);
|
2016-01-14 05:43:13 +08:00
|
|
|
else
|
2016-12-17 09:09:05 +08:00
|
|
|
setMinimumJumpTableEntries(std::numeric_limits<int>::max());
|
2015-12-19 04:19:30 +08:00
|
|
|
setOperationAction(ISD::BR_JT, MVT::Other, Expand);
|
2015-04-23 05:17:00 +08:00
|
|
|
|
|
|
|
// Hexagon has instructions for add/sub with carry. The problem with
|
|
|
|
// modeling these instructions is that they produce 2 results: Rdd and Px.
|
|
|
|
// To model the update of Px, we will have to use Defs[p0..p3] which will
|
|
|
|
// cause any predicate live range to spill. So, we pretend we dont't have
|
|
|
|
// these instructions.
|
|
|
|
setOperationAction(ISD::ADDE, MVT::i8, Expand);
|
2014-06-27 08:13:49 +08:00
|
|
|
setOperationAction(ISD::ADDE, MVT::i16, Expand);
|
|
|
|
setOperationAction(ISD::ADDE, MVT::i32, Expand);
|
|
|
|
setOperationAction(ISD::ADDE, MVT::i64, Expand);
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::SUBE, MVT::i8, Expand);
|
2014-06-27 08:13:49 +08:00
|
|
|
setOperationAction(ISD::SUBE, MVT::i16, Expand);
|
|
|
|
setOperationAction(ISD::SUBE, MVT::i32, Expand);
|
|
|
|
setOperationAction(ISD::SUBE, MVT::i64, Expand);
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::ADDC, MVT::i8, Expand);
|
2014-06-27 08:13:49 +08:00
|
|
|
setOperationAction(ISD::ADDC, MVT::i16, Expand);
|
|
|
|
setOperationAction(ISD::ADDC, MVT::i32, Expand);
|
|
|
|
setOperationAction(ISD::ADDC, MVT::i64, Expand);
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::SUBC, MVT::i8, Expand);
|
2014-06-27 08:13:49 +08:00
|
|
|
setOperationAction(ISD::SUBC, MVT::i16, Expand);
|
|
|
|
setOperationAction(ISD::SUBC, MVT::i32, Expand);
|
|
|
|
setOperationAction(ISD::SUBC, MVT::i64, Expand);
|
|
|
|
|
2015-04-14 04:37:01 +08:00
|
|
|
// Only add and sub that detect overflow are the saturating ones.
|
|
|
|
for (MVT VT : MVT::integer_valuetypes()) {
|
|
|
|
setOperationAction(ISD::UADDO, VT, Expand);
|
|
|
|
setOperationAction(ISD::SADDO, VT, Expand);
|
|
|
|
setOperationAction(ISD::USUBO, VT, Expand);
|
|
|
|
setOperationAction(ISD::SSUBO, VT, Expand);
|
|
|
|
}
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::CTLZ, MVT::i8, Promote);
|
|
|
|
setOperationAction(ISD::CTLZ, MVT::i16, Promote);
|
|
|
|
setOperationAction(ISD::CTTZ, MVT::i8, Promote);
|
|
|
|
setOperationAction(ISD::CTTZ, MVT::i16, Promote);
|
|
|
|
|
|
|
|
// In V5, popcount can count # of 1s in i64 but returns i32.
|
|
|
|
// On V4 it will be expanded (set later).
|
|
|
|
setOperationAction(ISD::CTPOP, MVT::i8, Promote);
|
|
|
|
setOperationAction(ISD::CTPOP, MVT::i16, Promote);
|
|
|
|
setOperationAction(ISD::CTPOP, MVT::i32, Promote);
|
2017-02-23 23:02:09 +08:00
|
|
|
setOperationAction(ISD::CTPOP, MVT::i64, Legal);
|
|
|
|
|
|
|
|
setOperationAction(ISD::BITREVERSE, MVT::i32, Legal);
|
|
|
|
setOperationAction(ISD::BITREVERSE, MVT::i64, Legal);
|
|
|
|
setOperationAction(ISD::BSWAP, MVT::i32, Legal);
|
|
|
|
setOperationAction(ISD::BSWAP, MVT::i64, Legal);
|
2017-05-31 01:47:51 +08:00
|
|
|
setOperationAction(ISD::MUL, MVT::i64, Legal);
|
2014-06-27 08:13:49 +08:00
|
|
|
|
2015-04-25 22:46:53 +08:00
|
|
|
for (unsigned IntExpOp :
|
2015-11-27 02:38:27 +08:00
|
|
|
{ ISD::SDIV, ISD::UDIV, ISD::SREM, ISD::UREM,
|
|
|
|
ISD::SDIVREM, ISD::UDIVREM, ISD::ROTL, ISD::ROTR,
|
2017-02-23 23:02:09 +08:00
|
|
|
ISD::SHL_PARTS, ISD::SRA_PARTS, ISD::SRL_PARTS,
|
2015-11-27 02:38:27 +08:00
|
|
|
ISD::SMUL_LOHI, ISD::UMUL_LOHI }) {
|
2015-04-25 22:46:53 +08:00
|
|
|
setOperationAction(IntExpOp, MVT::i32, Expand);
|
|
|
|
setOperationAction(IntExpOp, MVT::i64, Expand);
|
|
|
|
}
|
2014-06-27 08:13:49 +08:00
|
|
|
|
2015-04-25 22:46:53 +08:00
|
|
|
for (unsigned FPExpOp :
|
|
|
|
{ISD::FDIV, ISD::FREM, ISD::FSQRT, ISD::FSIN, ISD::FCOS, ISD::FSINCOS,
|
|
|
|
ISD::FPOW, ISD::FCOPYSIGN}) {
|
|
|
|
setOperationAction(FPExpOp, MVT::f32, Expand);
|
|
|
|
setOperationAction(FPExpOp, MVT::f64, Expand);
|
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
// No extending loads from i32.
|
|
|
|
for (MVT VT : MVT::integer_valuetypes()) {
|
|
|
|
setLoadExtAction(ISD::ZEXTLOAD, VT, MVT::i32, Expand);
|
|
|
|
setLoadExtAction(ISD::SEXTLOAD, VT, MVT::i32, Expand);
|
|
|
|
setLoadExtAction(ISD::EXTLOAD, VT, MVT::i32, Expand);
|
|
|
|
}
|
|
|
|
// Turn FP truncstore into trunc + store.
|
|
|
|
setTruncStoreAction(MVT::f64, MVT::f32, Expand);
|
2016-08-19 04:08:15 +08:00
|
|
|
// Turn FP extload into load/fpextend.
|
2015-04-23 05:17:00 +08:00
|
|
|
for (MVT VT : MVT::fp_valuetypes())
|
|
|
|
setLoadExtAction(ISD::EXTLOAD, VT, MVT::f32, Expand);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
// Expand BR_CC and SELECT_CC for all integer and fp types.
|
|
|
|
for (MVT VT : MVT::integer_valuetypes()) {
|
|
|
|
setOperationAction(ISD::BR_CC, VT, Expand);
|
|
|
|
setOperationAction(ISD::SELECT_CC, VT, Expand);
|
|
|
|
}
|
|
|
|
for (MVT VT : MVT::fp_valuetypes()) {
|
|
|
|
setOperationAction(ISD::BR_CC, VT, Expand);
|
|
|
|
setOperationAction(ISD::SELECT_CC, VT, Expand);
|
|
|
|
}
|
|
|
|
setOperationAction(ISD::BR_CC, MVT::Other, Expand);
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
//
|
|
|
|
// Handling of vector operations.
|
|
|
|
//
|
2011-12-13 05:14:40 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
promoteLdStType(MVT::v4i8, MVT::i32);
|
|
|
|
promoteLdStType(MVT::v2i16, MVT::i32);
|
|
|
|
promoteLdStType(MVT::v8i8, MVT::i64);
|
2017-07-17 23:45:45 +08:00
|
|
|
promoteLdStType(MVT::v4i16, MVT::i64);
|
2015-04-23 05:17:00 +08:00
|
|
|
promoteLdStType(MVT::v2i32, MVT::i64);
|
|
|
|
|
|
|
|
// Set the action for vector operations to "expand", then override it with
|
|
|
|
// either "custom" or "legal" for specific cases.
|
2015-10-18 13:15:34 +08:00
|
|
|
static const unsigned VectExpOps[] = {
|
2015-04-23 05:17:00 +08:00
|
|
|
// Integer arithmetic:
|
|
|
|
ISD::ADD, ISD::SUB, ISD::MUL, ISD::SDIV, ISD::UDIV,
|
|
|
|
ISD::SREM, ISD::UREM, ISD::SDIVREM, ISD::UDIVREM, ISD::ADDC,
|
|
|
|
ISD::SUBC, ISD::SADDO, ISD::UADDO, ISD::SSUBO, ISD::USUBO,
|
|
|
|
ISD::SMUL_LOHI, ISD::UMUL_LOHI,
|
|
|
|
// Logical/bit:
|
|
|
|
ISD::AND, ISD::OR, ISD::XOR, ISD::ROTL, ISD::ROTR,
|
2016-04-28 11:34:31 +08:00
|
|
|
ISD::CTPOP, ISD::CTLZ, ISD::CTTZ,
|
2015-04-23 05:17:00 +08:00
|
|
|
// Floating point arithmetic/math functions:
|
|
|
|
ISD::FADD, ISD::FSUB, ISD::FMUL, ISD::FMA, ISD::FDIV,
|
|
|
|
ISD::FREM, ISD::FNEG, ISD::FABS, ISD::FSQRT, ISD::FSIN,
|
2017-05-30 23:27:55 +08:00
|
|
|
ISD::FCOS, ISD::FPOW, ISD::FLOG, ISD::FLOG2,
|
2015-04-23 05:17:00 +08:00
|
|
|
ISD::FLOG10, ISD::FEXP, ISD::FEXP2, ISD::FCEIL, ISD::FTRUNC,
|
|
|
|
ISD::FRINT, ISD::FNEARBYINT, ISD::FROUND, ISD::FFLOOR,
|
|
|
|
ISD::FMINNUM, ISD::FMAXNUM, ISD::FSINCOS,
|
|
|
|
// Misc:
|
2016-10-27 22:30:16 +08:00
|
|
|
ISD::BR_CC, ISD::SELECT_CC, ISD::ConstantPool,
|
2015-04-23 05:17:00 +08:00
|
|
|
// Vector:
|
|
|
|
ISD::BUILD_VECTOR, ISD::SCALAR_TO_VECTOR,
|
|
|
|
ISD::EXTRACT_VECTOR_ELT, ISD::INSERT_VECTOR_ELT,
|
|
|
|
ISD::EXTRACT_SUBVECTOR, ISD::INSERT_SUBVECTOR,
|
|
|
|
ISD::CONCAT_VECTORS, ISD::VECTOR_SHUFFLE
|
|
|
|
};
|
|
|
|
|
|
|
|
for (MVT VT : MVT::vector_valuetypes()) {
|
2015-04-25 22:46:53 +08:00
|
|
|
for (unsigned VectExpOp : VectExpOps)
|
|
|
|
setOperationAction(VectExpOp, VT, Expand);
|
2015-04-23 05:17:00 +08:00
|
|
|
|
2016-09-09 01:42:14 +08:00
|
|
|
// Expand all extending loads and truncating stores:
|
2015-04-23 05:17:00 +08:00
|
|
|
for (MVT TargetVT : MVT::vector_valuetypes()) {
|
2016-09-09 01:42:14 +08:00
|
|
|
if (TargetVT == VT)
|
|
|
|
continue;
|
2015-04-23 05:17:00 +08:00
|
|
|
setLoadExtAction(ISD::EXTLOAD, TargetVT, VT, Expand);
|
2016-09-09 01:42:14 +08:00
|
|
|
setLoadExtAction(ISD::ZEXTLOAD, TargetVT, VT, Expand);
|
|
|
|
setLoadExtAction(ISD::SEXTLOAD, TargetVT, VT, Expand);
|
2015-04-23 05:17:00 +08:00
|
|
|
setTruncStoreAction(VT, TargetVT, Expand);
|
|
|
|
}
|
|
|
|
|
2016-10-27 22:30:16 +08:00
|
|
|
// Normalize all inputs to SELECT to be vectors of i32.
|
|
|
|
if (VT.getVectorElementType() != MVT::i32) {
|
|
|
|
MVT VT32 = MVT::getVectorVT(MVT::i32, VT.getSizeInBits()/32);
|
|
|
|
setOperationAction(ISD::SELECT, VT, Promote);
|
|
|
|
AddPromotedToType(ISD::SELECT, VT, VT32);
|
|
|
|
}
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::SRA, VT, Custom);
|
|
|
|
setOperationAction(ISD::SHL, VT, Custom);
|
|
|
|
setOperationAction(ISD::SRL, VT, Custom);
|
|
|
|
}
|
|
|
|
|
2017-10-21 03:33:12 +08:00
|
|
|
// Extending loads from (native) vectors of i8 into (native) vectors of i16
|
|
|
|
// are legal.
|
|
|
|
setLoadExtAction(ISD::EXTLOAD, MVT::v2i16, MVT::v2i8, Legal);
|
|
|
|
setLoadExtAction(ISD::ZEXTLOAD, MVT::v2i16, MVT::v2i8, Legal);
|
|
|
|
setLoadExtAction(ISD::SEXTLOAD, MVT::v2i16, MVT::v2i8, Legal);
|
|
|
|
setLoadExtAction(ISD::EXTLOAD, MVT::v4i16, MVT::v4i8, Legal);
|
|
|
|
setLoadExtAction(ISD::ZEXTLOAD, MVT::v4i16, MVT::v4i8, Legal);
|
|
|
|
setLoadExtAction(ISD::SEXTLOAD, MVT::v4i16, MVT::v4i8, Legal);
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
// Types natively supported:
|
2015-04-25 22:46:53 +08:00
|
|
|
for (MVT NativeVT : {MVT::v2i1, MVT::v4i1, MVT::v8i1, MVT::v32i1, MVT::v64i1,
|
|
|
|
MVT::v4i8, MVT::v8i8, MVT::v2i16, MVT::v4i16, MVT::v1i32,
|
|
|
|
MVT::v2i32, MVT::v1i64}) {
|
|
|
|
setOperationAction(ISD::BUILD_VECTOR, NativeVT, Custom);
|
|
|
|
setOperationAction(ISD::EXTRACT_VECTOR_ELT, NativeVT, Custom);
|
|
|
|
setOperationAction(ISD::INSERT_VECTOR_ELT, NativeVT, Custom);
|
|
|
|
setOperationAction(ISD::EXTRACT_SUBVECTOR, NativeVT, Custom);
|
|
|
|
setOperationAction(ISD::INSERT_SUBVECTOR, NativeVT, Custom);
|
|
|
|
setOperationAction(ISD::CONCAT_VECTORS, NativeVT, Custom);
|
|
|
|
|
|
|
|
setOperationAction(ISD::ADD, NativeVT, Legal);
|
|
|
|
setOperationAction(ISD::SUB, NativeVT, Legal);
|
|
|
|
setOperationAction(ISD::MUL, NativeVT, Legal);
|
|
|
|
setOperationAction(ISD::AND, NativeVT, Legal);
|
|
|
|
setOperationAction(ISD::OR, NativeVT, Legal);
|
|
|
|
setOperationAction(ISD::XOR, NativeVT, Legal);
|
2015-04-23 05:17:00 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
setOperationAction(ISD::SETCC, MVT::v2i16, Custom);
|
|
|
|
setOperationAction(ISD::VSELECT, MVT::v2i16, Custom);
|
|
|
|
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v4i16, Custom);
|
|
|
|
setOperationAction(ISD::VECTOR_SHUFFLE, MVT::v8i8, Custom);
|
2016-09-14 05:16:07 +08:00
|
|
|
|
2017-12-07 00:40:37 +08:00
|
|
|
auto setPromoteTo = [this] (unsigned Opc, MVT FromTy, MVT ToTy) {
|
|
|
|
setOperationAction(Opc, FromTy, Promote);
|
|
|
|
AddPromotedToType(Opc, FromTy, ToTy);
|
|
|
|
};
|
|
|
|
|
2017-10-19 01:45:22 +08:00
|
|
|
if (Subtarget.useHVXOps()) {
|
2017-12-07 00:40:37 +08:00
|
|
|
bool Use64b = Subtarget.useHVX64BOps();
|
|
|
|
ArrayRef<MVT> LegalV = Use64b ? LegalV64 : LegalV128;
|
|
|
|
ArrayRef<MVT> LegalW = Use64b ? LegalW64 : LegalW128;
|
2017-12-08 01:37:28 +08:00
|
|
|
MVT ByteV = Use64b ? MVT::v64i8 : MVT::v128i8;
|
|
|
|
MVT ByteW = Use64b ? MVT::v128i8 : MVT::v256i8;
|
2017-12-07 00:40:37 +08:00
|
|
|
|
|
|
|
setOperationAction(ISD::VECTOR_SHUFFLE, ByteV, Legal);
|
|
|
|
setOperationAction(ISD::VECTOR_SHUFFLE, ByteW, Legal);
|
|
|
|
setOperationAction(ISD::CONCAT_VECTORS, ByteW, Legal);
|
2017-12-08 01:37:28 +08:00
|
|
|
setOperationAction(ISD::AND, ByteV, Legal);
|
2017-12-07 00:40:37 +08:00
|
|
|
setOperationAction(ISD::OR, ByteV, Legal);
|
2017-12-08 01:37:28 +08:00
|
|
|
setOperationAction(ISD::XOR, ByteV, Legal);
|
2017-12-07 00:40:37 +08:00
|
|
|
|
|
|
|
for (MVT T : LegalV) {
|
|
|
|
setIndexedLoadAction(ISD::POST_INC, T, Legal);
|
|
|
|
setIndexedStoreAction(ISD::POST_INC, T, Legal);
|
|
|
|
|
2017-12-15 05:28:48 +08:00
|
|
|
setOperationAction(ISD::ADD, T, Legal);
|
|
|
|
setOperationAction(ISD::SUB, T, Legal);
|
|
|
|
setOperationAction(ISD::VSELECT, T, Legal);
|
2017-12-19 02:32:27 +08:00
|
|
|
if (T != ByteV) {
|
|
|
|
setOperationAction(ISD::SIGN_EXTEND_VECTOR_INREG, T, Legal);
|
|
|
|
setOperationAction(ISD::ZERO_EXTEND_VECTOR_INREG, T, Legal);
|
|
|
|
}
|
2017-12-08 01:37:28 +08:00
|
|
|
|
2017-12-15 05:28:48 +08:00
|
|
|
setOperationAction(ISD::MUL, T, Custom);
|
|
|
|
setOperationAction(ISD::SETCC, T, Custom);
|
2017-12-07 00:40:37 +08:00
|
|
|
setOperationAction(ISD::BUILD_VECTOR, T, Custom);
|
|
|
|
setOperationAction(ISD::INSERT_SUBVECTOR, T, Custom);
|
|
|
|
setOperationAction(ISD::INSERT_VECTOR_ELT, T, Custom);
|
|
|
|
setOperationAction(ISD::EXTRACT_SUBVECTOR, T, Custom);
|
|
|
|
setOperationAction(ISD::EXTRACT_VECTOR_ELT, T, Custom);
|
2017-12-19 02:41:52 +08:00
|
|
|
if (T != ByteV)
|
|
|
|
setOperationAction(ISD::ANY_EXTEND_VECTOR_INREG, T, Custom);
|
2017-12-07 00:40:37 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
for (MVT T : LegalV) {
|
|
|
|
if (T == ByteV)
|
|
|
|
continue;
|
|
|
|
// Promote all shuffles and concats to operate on vectors of bytes.
|
|
|
|
setPromoteTo(ISD::VECTOR_SHUFFLE, T, ByteV);
|
|
|
|
setPromoteTo(ISD::CONCAT_VECTORS, T, ByteV);
|
2017-12-08 01:37:28 +08:00
|
|
|
setPromoteTo(ISD::AND, T, ByteV);
|
2017-12-07 00:40:37 +08:00
|
|
|
setPromoteTo(ISD::OR, T, ByteV);
|
2017-12-08 01:37:28 +08:00
|
|
|
setPromoteTo(ISD::XOR, T, ByteV);
|
2017-12-07 00:40:37 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
for (MVT T : LegalW) {
|
|
|
|
if (T == ByteW)
|
|
|
|
continue;
|
|
|
|
// Promote all shuffles and concats to operate on vectors of bytes.
|
|
|
|
setPromoteTo(ISD::VECTOR_SHUFFLE, T, ByteW);
|
|
|
|
setPromoteTo(ISD::CONCAT_VECTORS, T, ByteW);
|
2015-11-27 02:38:27 +08:00
|
|
|
}
|
|
|
|
}
|
2017-12-07 00:40:37 +08:00
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
// Subtarget-specific operation actions.
|
|
|
|
//
|
|
|
|
if (Subtarget.hasV5TOps()) {
|
|
|
|
setOperationAction(ISD::FMA, MVT::f64, Expand);
|
|
|
|
setOperationAction(ISD::FADD, MVT::f64, Expand);
|
|
|
|
setOperationAction(ISD::FSUB, MVT::f64, Expand);
|
|
|
|
setOperationAction(ISD::FMUL, MVT::f64, Expand);
|
|
|
|
|
2016-08-19 21:34:31 +08:00
|
|
|
setOperationAction(ISD::FMINNUM, MVT::f32, Legal);
|
|
|
|
setOperationAction(ISD::FMAXNUM, MVT::f32, Legal);
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
setOperationAction(ISD::FP_TO_UINT, MVT::i1, Promote);
|
|
|
|
setOperationAction(ISD::FP_TO_UINT, MVT::i8, Promote);
|
|
|
|
setOperationAction(ISD::FP_TO_UINT, MVT::i16, Promote);
|
|
|
|
setOperationAction(ISD::FP_TO_SINT, MVT::i1, Promote);
|
|
|
|
setOperationAction(ISD::FP_TO_SINT, MVT::i8, Promote);
|
|
|
|
setOperationAction(ISD::FP_TO_SINT, MVT::i16, Promote);
|
|
|
|
setOperationAction(ISD::UINT_TO_FP, MVT::i1, Promote);
|
|
|
|
setOperationAction(ISD::UINT_TO_FP, MVT::i8, Promote);
|
|
|
|
setOperationAction(ISD::UINT_TO_FP, MVT::i16, Promote);
|
|
|
|
setOperationAction(ISD::SINT_TO_FP, MVT::i1, Promote);
|
|
|
|
setOperationAction(ISD::SINT_TO_FP, MVT::i8, Promote);
|
|
|
|
setOperationAction(ISD::SINT_TO_FP, MVT::i16, Promote);
|
|
|
|
} else { // V4
|
|
|
|
setOperationAction(ISD::SINT_TO_FP, MVT::i32, Expand);
|
|
|
|
setOperationAction(ISD::SINT_TO_FP, MVT::i64, Expand);
|
|
|
|
setOperationAction(ISD::UINT_TO_FP, MVT::i32, Expand);
|
|
|
|
setOperationAction(ISD::UINT_TO_FP, MVT::i64, Expand);
|
|
|
|
setOperationAction(ISD::FP_TO_SINT, MVT::f64, Expand);
|
|
|
|
setOperationAction(ISD::FP_TO_SINT, MVT::f32, Expand);
|
|
|
|
setOperationAction(ISD::FP_EXTEND, MVT::f32, Expand);
|
|
|
|
setOperationAction(ISD::FP_ROUND, MVT::f64, Expand);
|
|
|
|
setCondCodeAction(ISD::SETUNE, MVT::f64, Expand);
|
|
|
|
|
|
|
|
setOperationAction(ISD::CTPOP, MVT::i8, Expand);
|
|
|
|
setOperationAction(ISD::CTPOP, MVT::i16, Expand);
|
|
|
|
setOperationAction(ISD::CTPOP, MVT::i32, Expand);
|
|
|
|
setOperationAction(ISD::CTPOP, MVT::i64, Expand);
|
|
|
|
|
|
|
|
// Expand these operations for both f32 and f64:
|
2015-04-25 22:46:53 +08:00
|
|
|
for (unsigned FPExpOpV4 :
|
|
|
|
{ISD::FADD, ISD::FSUB, ISD::FMUL, ISD::FABS, ISD::FNEG, ISD::FMA}) {
|
|
|
|
setOperationAction(FPExpOpV4, MVT::f32, Expand);
|
|
|
|
setOperationAction(FPExpOpV4, MVT::f64, Expand);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (ISD::CondCode FPExpCCV4 :
|
|
|
|
{ISD::SETOEQ, ISD::SETOGT, ISD::SETOLT, ISD::SETOGE, ISD::SETOLE,
|
2015-11-27 02:38:27 +08:00
|
|
|
ISD::SETUO, ISD::SETO}) {
|
2015-04-25 22:46:53 +08:00
|
|
|
setCondCodeAction(FPExpCCV4, MVT::f32, Expand);
|
|
|
|
setCondCodeAction(FPExpCCV4, MVT::f64, Expand);
|
2015-04-23 05:17:00 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Handling of indexed loads/stores: default is "expand".
|
|
|
|
//
|
2016-07-27 04:30:30 +08:00
|
|
|
for (MVT VT : {MVT::i8, MVT::i16, MVT::i32, MVT::i64}) {
|
|
|
|
setIndexedLoadAction(ISD::POST_INC, VT, Legal);
|
|
|
|
setIndexedStoreAction(ISD::POST_INC, VT, Legal);
|
2015-04-23 05:17:00 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
computeRegisterProperties(&HRI);
|
|
|
|
|
|
|
|
//
|
|
|
|
// Library calls for unsupported operations
|
|
|
|
//
|
|
|
|
bool FastMath = EnableFastMath;
|
|
|
|
|
2015-04-25 22:46:46 +08:00
|
|
|
setLibcallName(RTLIB::SDIV_I32, "__hexagon_divsi3");
|
|
|
|
setLibcallName(RTLIB::SDIV_I64, "__hexagon_divdi3");
|
|
|
|
setLibcallName(RTLIB::UDIV_I32, "__hexagon_udivsi3");
|
|
|
|
setLibcallName(RTLIB::UDIV_I64, "__hexagon_udivdi3");
|
|
|
|
setLibcallName(RTLIB::SREM_I32, "__hexagon_modsi3");
|
|
|
|
setLibcallName(RTLIB::SREM_I64, "__hexagon_moddi3");
|
|
|
|
setLibcallName(RTLIB::UREM_I32, "__hexagon_umodsi3");
|
|
|
|
setLibcallName(RTLIB::UREM_I64, "__hexagon_umoddi3");
|
|
|
|
|
|
|
|
setLibcallName(RTLIB::SINTTOFP_I128_F64, "__hexagon_floattidf");
|
|
|
|
setLibcallName(RTLIB::SINTTOFP_I128_F32, "__hexagon_floattisf");
|
|
|
|
setLibcallName(RTLIB::FPTOUINT_F32_I128, "__hexagon_fixunssfti");
|
|
|
|
setLibcallName(RTLIB::FPTOUINT_F64_I128, "__hexagon_fixunsdfti");
|
|
|
|
setLibcallName(RTLIB::FPTOSINT_F32_I128, "__hexagon_fixsfti");
|
|
|
|
setLibcallName(RTLIB::FPTOSINT_F64_I128, "__hexagon_fixdfti");
|
2015-04-23 05:17:00 +08:00
|
|
|
|
|
|
|
if (IsV4) {
|
|
|
|
// Handle single-precision floating point operations on V4.
|
2015-04-25 22:46:46 +08:00
|
|
|
if (FastMath) {
|
|
|
|
setLibcallName(RTLIB::ADD_F32, "__hexagon_fast_addsf3");
|
|
|
|
setLibcallName(RTLIB::SUB_F32, "__hexagon_fast_subsf3");
|
|
|
|
setLibcallName(RTLIB::MUL_F32, "__hexagon_fast_mulsf3");
|
|
|
|
setLibcallName(RTLIB::OGT_F32, "__hexagon_fast_gtsf2");
|
|
|
|
setLibcallName(RTLIB::OLT_F32, "__hexagon_fast_ltsf2");
|
|
|
|
// Double-precision compares.
|
|
|
|
setLibcallName(RTLIB::OGT_F64, "__hexagon_fast_gtdf2");
|
|
|
|
setLibcallName(RTLIB::OLT_F64, "__hexagon_fast_ltdf2");
|
|
|
|
} else {
|
|
|
|
setLibcallName(RTLIB::ADD_F32, "__hexagon_addsf3");
|
|
|
|
setLibcallName(RTLIB::SUB_F32, "__hexagon_subsf3");
|
|
|
|
setLibcallName(RTLIB::MUL_F32, "__hexagon_mulsf3");
|
|
|
|
setLibcallName(RTLIB::OGT_F32, "__hexagon_gtsf2");
|
|
|
|
setLibcallName(RTLIB::OLT_F32, "__hexagon_ltsf2");
|
|
|
|
// Double-precision compares.
|
|
|
|
setLibcallName(RTLIB::OGT_F64, "__hexagon_gtdf2");
|
|
|
|
setLibcallName(RTLIB::OLT_F64, "__hexagon_ltdf2");
|
|
|
|
}
|
2015-04-23 05:17:00 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// This is the only fast library function for sqrtd.
|
|
|
|
if (FastMath)
|
2015-04-25 22:46:46 +08:00
|
|
|
setLibcallName(RTLIB::SQRT_F64, "__hexagon_fast2_sqrtdf2");
|
2015-04-23 05:17:00 +08:00
|
|
|
|
2015-04-25 22:46:46 +08:00
|
|
|
// Prefix is: nothing for "slow-math",
|
|
|
|
// "fast2_" for V4 fast-math and V5+ fast-math double-precision
|
2015-04-23 05:17:00 +08:00
|
|
|
// (actually, keep fast-math and fast-math2 separate for now)
|
2015-04-25 22:46:46 +08:00
|
|
|
if (FastMath) {
|
|
|
|
setLibcallName(RTLIB::ADD_F64, "__hexagon_fast_adddf3");
|
|
|
|
setLibcallName(RTLIB::SUB_F64, "__hexagon_fast_subdf3");
|
|
|
|
setLibcallName(RTLIB::MUL_F64, "__hexagon_fast_muldf3");
|
|
|
|
setLibcallName(RTLIB::DIV_F64, "__hexagon_fast_divdf3");
|
|
|
|
// Calling __hexagon_fast2_divsf3 with fast-math on V5 (ok).
|
|
|
|
setLibcallName(RTLIB::DIV_F32, "__hexagon_fast_divsf3");
|
|
|
|
} else {
|
|
|
|
setLibcallName(RTLIB::ADD_F64, "__hexagon_adddf3");
|
|
|
|
setLibcallName(RTLIB::SUB_F64, "__hexagon_subdf3");
|
|
|
|
setLibcallName(RTLIB::MUL_F64, "__hexagon_muldf3");
|
|
|
|
setLibcallName(RTLIB::DIV_F64, "__hexagon_divdf3");
|
|
|
|
setLibcallName(RTLIB::DIV_F32, "__hexagon_divsf3");
|
|
|
|
}
|
2015-04-23 05:17:00 +08:00
|
|
|
|
|
|
|
if (Subtarget.hasV5TOps()) {
|
|
|
|
if (FastMath)
|
2015-04-25 22:46:46 +08:00
|
|
|
setLibcallName(RTLIB::SQRT_F32, "__hexagon_fast2_sqrtf");
|
2015-04-23 05:17:00 +08:00
|
|
|
else
|
2015-04-25 22:46:46 +08:00
|
|
|
setLibcallName(RTLIB::SQRT_F32, "__hexagon_sqrtf");
|
2015-04-23 05:17:00 +08:00
|
|
|
} else {
|
|
|
|
// V4
|
2015-04-25 22:46:46 +08:00
|
|
|
setLibcallName(RTLIB::SINTTOFP_I32_F32, "__hexagon_floatsisf");
|
|
|
|
setLibcallName(RTLIB::SINTTOFP_I32_F64, "__hexagon_floatsidf");
|
|
|
|
setLibcallName(RTLIB::SINTTOFP_I64_F32, "__hexagon_floatdisf");
|
|
|
|
setLibcallName(RTLIB::SINTTOFP_I64_F64, "__hexagon_floatdidf");
|
|
|
|
setLibcallName(RTLIB::UINTTOFP_I32_F32, "__hexagon_floatunsisf");
|
|
|
|
setLibcallName(RTLIB::UINTTOFP_I32_F64, "__hexagon_floatunsidf");
|
|
|
|
setLibcallName(RTLIB::UINTTOFP_I64_F32, "__hexagon_floatundisf");
|
|
|
|
setLibcallName(RTLIB::UINTTOFP_I64_F64, "__hexagon_floatundidf");
|
|
|
|
setLibcallName(RTLIB::FPTOUINT_F32_I32, "__hexagon_fixunssfsi");
|
|
|
|
setLibcallName(RTLIB::FPTOUINT_F32_I64, "__hexagon_fixunssfdi");
|
|
|
|
setLibcallName(RTLIB::FPTOUINT_F64_I32, "__hexagon_fixunsdfsi");
|
|
|
|
setLibcallName(RTLIB::FPTOUINT_F64_I64, "__hexagon_fixunsdfdi");
|
|
|
|
setLibcallName(RTLIB::FPTOSINT_F32_I32, "__hexagon_fixsfsi");
|
|
|
|
setLibcallName(RTLIB::FPTOSINT_F32_I64, "__hexagon_fixsfdi");
|
|
|
|
setLibcallName(RTLIB::FPTOSINT_F64_I32, "__hexagon_fixdfsi");
|
|
|
|
setLibcallName(RTLIB::FPTOSINT_F64_I64, "__hexagon_fixdfdi");
|
|
|
|
setLibcallName(RTLIB::FPEXT_F32_F64, "__hexagon_extendsfdf2");
|
|
|
|
setLibcallName(RTLIB::FPROUND_F64_F32, "__hexagon_truncdfsf2");
|
|
|
|
setLibcallName(RTLIB::OEQ_F32, "__hexagon_eqsf2");
|
|
|
|
setLibcallName(RTLIB::OEQ_F64, "__hexagon_eqdf2");
|
|
|
|
setLibcallName(RTLIB::OGE_F32, "__hexagon_gesf2");
|
|
|
|
setLibcallName(RTLIB::OGE_F64, "__hexagon_gedf2");
|
|
|
|
setLibcallName(RTLIB::OLE_F32, "__hexagon_lesf2");
|
|
|
|
setLibcallName(RTLIB::OLE_F64, "__hexagon_ledf2");
|
|
|
|
setLibcallName(RTLIB::UNE_F32, "__hexagon_nesf2");
|
|
|
|
setLibcallName(RTLIB::UNE_F64, "__hexagon_nedf2");
|
|
|
|
setLibcallName(RTLIB::UO_F32, "__hexagon_unordsf2");
|
|
|
|
setLibcallName(RTLIB::UO_F64, "__hexagon_unorddf2");
|
|
|
|
setLibcallName(RTLIB::O_F32, "__hexagon_unordsf2");
|
|
|
|
setLibcallName(RTLIB::O_F64, "__hexagon_unorddf2");
|
2015-04-23 05:17:00 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// These cause problems when the shift amount is non-constant.
|
|
|
|
setLibcallName(RTLIB::SHL_I128, nullptr);
|
|
|
|
setLibcallName(RTLIB::SRL_I128, nullptr);
|
|
|
|
setLibcallName(RTLIB::SRA_I128, nullptr);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
const char* HexagonTargetLowering::getTargetNodeName(unsigned Opcode) const {
|
2015-05-08 05:33:59 +08:00
|
|
|
switch ((HexagonISD::NodeType)Opcode) {
|
2015-04-23 05:17:00 +08:00
|
|
|
case HexagonISD::ALLOCA: return "HexagonISD::ALLOCA";
|
|
|
|
case HexagonISD::AT_GOT: return "HexagonISD::AT_GOT";
|
|
|
|
case HexagonISD::AT_PCREL: return "HexagonISD::AT_PCREL";
|
|
|
|
case HexagonISD::BARRIER: return "HexagonISD::BARRIER";
|
2016-08-12 19:12:02 +08:00
|
|
|
case HexagonISD::CALL: return "HexagonISD::CALL";
|
|
|
|
case HexagonISD::CALLnr: return "HexagonISD::CALLnr";
|
2015-04-23 05:17:00 +08:00
|
|
|
case HexagonISD::CALLR: return "HexagonISD::CALLR";
|
|
|
|
case HexagonISD::COMBINE: return "HexagonISD::COMBINE";
|
|
|
|
case HexagonISD::CONST32_GP: return "HexagonISD::CONST32_GP";
|
|
|
|
case HexagonISD::CONST32: return "HexagonISD::CONST32";
|
|
|
|
case HexagonISD::CP: return "HexagonISD::CP";
|
|
|
|
case HexagonISD::DCFETCH: return "HexagonISD::DCFETCH";
|
|
|
|
case HexagonISD::EH_RETURN: return "HexagonISD::EH_RETURN";
|
|
|
|
case HexagonISD::EXTRACTU: return "HexagonISD::EXTRACTU";
|
|
|
|
case HexagonISD::EXTRACTURP: return "HexagonISD::EXTRACTURP";
|
|
|
|
case HexagonISD::INSERT: return "HexagonISD::INSERT";
|
|
|
|
case HexagonISD::INSERTRP: return "HexagonISD::INSERTRP";
|
|
|
|
case HexagonISD::JT: return "HexagonISD::JT";
|
|
|
|
case HexagonISD::RET_FLAG: return "HexagonISD::RET_FLAG";
|
|
|
|
case HexagonISD::TC_RETURN: return "HexagonISD::TC_RETURN";
|
2015-12-04 00:47:20 +08:00
|
|
|
case HexagonISD::VCOMBINE: return "HexagonISD::VCOMBINE";
|
2017-07-15 03:02:32 +08:00
|
|
|
case HexagonISD::VPACKE: return "HexagonISD::VPACKE";
|
|
|
|
case HexagonISD::VPACKO: return "HexagonISD::VPACKO";
|
2017-07-11 04:16:44 +08:00
|
|
|
case HexagonISD::VASL: return "HexagonISD::VASL";
|
|
|
|
case HexagonISD::VASR: return "HexagonISD::VASR";
|
|
|
|
case HexagonISD::VLSR: return "HexagonISD::VLSR";
|
|
|
|
case HexagonISD::VSPLAT: return "HexagonISD::VSPLAT";
|
2017-12-07 00:40:37 +08:00
|
|
|
case HexagonISD::VEXTRACTW: return "HexagonISD::VEXTRACTW";
|
|
|
|
case HexagonISD::VINSERTW0: return "HexagonISD::VINSERTW0";
|
|
|
|
case HexagonISD::VROR: return "HexagonISD::VROR";
|
2017-02-23 06:28:47 +08:00
|
|
|
case HexagonISD::READCYCLE: return "HexagonISD::READCYCLE";
|
2015-05-08 05:33:59 +08:00
|
|
|
case HexagonISD::OP_END: break;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
2015-05-08 05:33:59 +08:00
|
|
|
return nullptr;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2017-12-12 02:57:54 +08:00
|
|
|
/// Given an intrinsic, checks if on the target the intrinsic will need to map
|
|
|
|
/// to a MemIntrinsicNode (touches memory). If this is the case, it returns
|
|
|
|
/// true and store the intrinsic information into the IntrinsicInfo that was
|
|
|
|
/// passed to the function.
|
|
|
|
bool HexagonTargetLowering::getTgtMemIntrinsic(IntrinsicInfo &Info,
|
|
|
|
const CallInst &I,
|
2017-12-15 06:34:10 +08:00
|
|
|
MachineFunction &MF,
|
2017-12-12 02:57:54 +08:00
|
|
|
unsigned Intrinsic) const {
|
|
|
|
switch (Intrinsic) {
|
|
|
|
case Intrinsic::hexagon_V6_vgathermw:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermw_128B:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermh:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermh_128B:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermhw:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermhw_128B:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermwq:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermwq_128B:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermhq:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermhq_128B:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermhwq:
|
|
|
|
case Intrinsic::hexagon_V6_vgathermhwq_128B: {
|
|
|
|
const Module &M = *I.getParent()->getParent()->getParent();
|
|
|
|
Info.opc = ISD::INTRINSIC_W_CHAIN;
|
|
|
|
Type *VecTy = I.getArgOperand(1)->getType();
|
|
|
|
Info.memVT = MVT::getVT(VecTy);
|
|
|
|
Info.ptrVal = I.getArgOperand(0);
|
|
|
|
Info.offset = 0;
|
|
|
|
Info.align = M.getDataLayout().getTypeAllocSizeInBits(VecTy) / 8;
|
2017-12-15 05:39:51 +08:00
|
|
|
Info.flags = MachineMemOperand::MOLoad |
|
|
|
|
MachineMemOperand::MOStore |
|
|
|
|
MachineMemOperand::MOVolatile;
|
2017-12-12 02:57:54 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2015-04-23 05:17:00 +08:00
|
|
|
bool HexagonTargetLowering::isTruncateFree(Type *Ty1, Type *Ty2) const {
|
2011-12-13 05:14:40 +08:00
|
|
|
EVT MTy1 = EVT::getEVT(Ty1);
|
|
|
|
EVT MTy2 = EVT::getEVT(Ty2);
|
2015-04-23 05:17:00 +08:00
|
|
|
if (!MTy1.isSimple() || !MTy2.isSimple())
|
2011-12-13 05:14:40 +08:00
|
|
|
return false;
|
2015-04-23 05:17:00 +08:00
|
|
|
return (MTy1.getSimpleVT() == MVT::i64) && (MTy2.getSimpleVT() == MVT::i32);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bool HexagonTargetLowering::isTruncateFree(EVT VT1, EVT VT2) const {
|
2015-04-23 05:17:00 +08:00
|
|
|
if (!VT1.isSimple() || !VT2.isSimple())
|
2011-12-13 05:14:40 +08:00
|
|
|
return false;
|
2015-04-23 05:17:00 +08:00
|
|
|
return (VT1.getSimpleVT() == MVT::i64) && (VT2.getSimpleVT() == MVT::i32);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2016-08-19 21:34:31 +08:00
|
|
|
bool HexagonTargetLowering::isFMAFasterThanFMulAndFAdd(EVT VT) const {
|
|
|
|
return isOperationLegalOrCustom(ISD::FMA, VT);
|
|
|
|
}
|
|
|
|
|
2015-03-20 00:33:08 +08:00
|
|
|
// Should we expand the build vector with shuffles?
|
2016-09-14 05:16:07 +08:00
|
|
|
bool HexagonTargetLowering::shouldExpandBuildVectorWithShuffles(EVT VT,
|
|
|
|
unsigned DefinedValues) const {
|
2017-12-07 00:40:37 +08:00
|
|
|
return false;
|
2016-07-30 00:44:27 +08:00
|
|
|
}
|
|
|
|
|
2017-07-26 16:06:58 +08:00
|
|
|
bool HexagonTargetLowering::isShuffleMaskLegal(ArrayRef<int> Mask,
|
|
|
|
EVT VT) const {
|
2016-09-14 05:16:07 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2017-12-19 02:21:01 +08:00
|
|
|
TargetLoweringBase::LegalizeTypeAction
|
|
|
|
HexagonTargetLowering::getPreferredVectorAction(EVT VT) const {
|
|
|
|
if (Subtarget.useHVXOps()) {
|
|
|
|
// If the size of VT is at least half of the vector length,
|
|
|
|
// widen the vector. Note: the threshold was not selected in
|
|
|
|
// any scientific way.
|
|
|
|
if (VT.getSizeInBits() >= Subtarget.getVectorLength()*8/2)
|
|
|
|
return TargetLoweringBase::TypeWidenVector;
|
|
|
|
}
|
|
|
|
return TargetLowering::getPreferredVectorAction(VT);
|
|
|
|
}
|
|
|
|
|
2016-07-30 00:44:27 +08:00
|
|
|
// Lower a vector shuffle (V1, V2, V3). V1 and V2 are the two vectors
|
|
|
|
// to select data from, V3 is the permutation.
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerVECTOR_SHUFFLE(SDValue Op, SelectionDAG &DAG)
|
|
|
|
const {
|
2015-03-20 00:33:08 +08:00
|
|
|
const ShuffleVectorSDNode *SVN = cast<ShuffleVectorSDNode>(Op);
|
|
|
|
SDValue V1 = Op.getOperand(0);
|
|
|
|
SDValue V2 = Op.getOperand(1);
|
|
|
|
SDLoc dl(Op);
|
|
|
|
EVT VT = Op.getValueType();
|
|
|
|
|
2016-03-15 01:28:46 +08:00
|
|
|
if (V2.isUndef())
|
2015-03-20 00:33:08 +08:00
|
|
|
V2 = V1;
|
|
|
|
|
|
|
|
if (SVN->isSplat()) {
|
|
|
|
int Lane = SVN->getSplatIndex();
|
|
|
|
if (Lane == -1) Lane = 0;
|
|
|
|
|
|
|
|
// Test if V1 is a SCALAR_TO_VECTOR.
|
|
|
|
if (Lane == 0 && V1.getOpcode() == ISD::SCALAR_TO_VECTOR)
|
2017-07-14 02:17:58 +08:00
|
|
|
return DAG.getNode(HexagonISD::VSPLAT, dl, VT, V1.getOperand(0));
|
2015-03-20 00:33:08 +08:00
|
|
|
|
|
|
|
// Test if V1 is a BUILD_VECTOR which is equivalent to a SCALAR_TO_VECTOR
|
|
|
|
// (and probably will turn into a SCALAR_TO_VECTOR once legalization
|
|
|
|
// reaches it).
|
|
|
|
if (Lane == 0 && V1.getOpcode() == ISD::BUILD_VECTOR &&
|
|
|
|
!isa<ConstantSDNode>(V1.getOperand(0))) {
|
|
|
|
bool IsScalarToVector = true;
|
2016-08-02 04:31:50 +08:00
|
|
|
for (unsigned i = 1, e = V1.getNumOperands(); i != e; ++i) {
|
2016-03-15 02:09:43 +08:00
|
|
|
if (!V1.getOperand(i).isUndef()) {
|
2015-03-20 00:33:08 +08:00
|
|
|
IsScalarToVector = false;
|
|
|
|
break;
|
|
|
|
}
|
2016-08-02 04:31:50 +08:00
|
|
|
}
|
2015-03-20 00:33:08 +08:00
|
|
|
if (IsScalarToVector)
|
2017-07-14 02:17:58 +08:00
|
|
|
return DAG.getNode(HexagonISD::VSPLAT, dl, VT, V1.getOperand(0));
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
2017-07-14 02:17:58 +08:00
|
|
|
return DAG.getNode(HexagonISD::VSPLAT, dl, VT,
|
|
|
|
DAG.getConstant(Lane, dl, MVT::i32));
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// FIXME: We need to support more general vector shuffles. See
|
|
|
|
// below the comment from the ARM backend that deals in the general
|
|
|
|
// case with the vector shuffles. For now, let expand handle these.
|
|
|
|
return SDValue();
|
|
|
|
|
|
|
|
// If the shuffle is not directly supported and it has 4 elements, use
|
|
|
|
// the PerfectShuffle-generated table to synthesize it from other shuffles.
|
|
|
|
}
|
|
|
|
|
|
|
|
// If BUILD_VECTOR has same base element repeated several times,
|
|
|
|
// report true.
|
|
|
|
static bool isCommonSplatElement(BuildVectorSDNode *BVN) {
|
|
|
|
unsigned NElts = BVN->getNumOperands();
|
|
|
|
SDValue V0 = BVN->getOperand(0);
|
|
|
|
|
|
|
|
for (unsigned i = 1, e = NElts; i != e; ++i) {
|
|
|
|
if (BVN->getOperand(i) != V0)
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2016-07-30 00:44:27 +08:00
|
|
|
// Lower a vector shift. Try to convert
|
2015-03-20 00:33:08 +08:00
|
|
|
// <VT> = SHL/SRA/SRL <VT> by <VT> to Hexagon specific
|
|
|
|
// <VT> = SHL/SRA/SRL <VT> by <IT/i32>.
|
2016-07-30 00:44:27 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerVECTOR_SHIFT(SDValue Op, SelectionDAG &DAG) const {
|
2016-12-17 09:09:05 +08:00
|
|
|
BuildVectorSDNode *BVN = nullptr;
|
2015-03-20 00:33:08 +08:00
|
|
|
SDValue V1 = Op.getOperand(0);
|
|
|
|
SDValue V2 = Op.getOperand(1);
|
|
|
|
SDValue V3;
|
|
|
|
SDLoc dl(Op);
|
|
|
|
EVT VT = Op.getValueType();
|
|
|
|
|
|
|
|
if ((BVN = dyn_cast<BuildVectorSDNode>(V1.getNode())) &&
|
|
|
|
isCommonSplatElement(BVN))
|
|
|
|
V3 = V2;
|
|
|
|
else if ((BVN = dyn_cast<BuildVectorSDNode>(V2.getNode())) &&
|
|
|
|
isCommonSplatElement(BVN))
|
|
|
|
V3 = V1;
|
|
|
|
else
|
|
|
|
return SDValue();
|
|
|
|
|
|
|
|
SDValue CommonSplat = BVN->getOperand(0);
|
|
|
|
SDValue Result;
|
|
|
|
|
|
|
|
if (VT.getSimpleVT() == MVT::v4i16) {
|
|
|
|
switch (Op.getOpcode()) {
|
|
|
|
case ISD::SRA:
|
2017-07-11 04:16:44 +08:00
|
|
|
Result = DAG.getNode(HexagonISD::VASR, dl, VT, V3, CommonSplat);
|
2015-03-20 00:33:08 +08:00
|
|
|
break;
|
|
|
|
case ISD::SHL:
|
2017-07-11 04:16:44 +08:00
|
|
|
Result = DAG.getNode(HexagonISD::VASL, dl, VT, V3, CommonSplat);
|
2015-03-20 00:33:08 +08:00
|
|
|
break;
|
|
|
|
case ISD::SRL:
|
2017-07-11 04:16:44 +08:00
|
|
|
Result = DAG.getNode(HexagonISD::VLSR, dl, VT, V3, CommonSplat);
|
2015-03-20 00:33:08 +08:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return SDValue();
|
|
|
|
}
|
|
|
|
} else if (VT.getSimpleVT() == MVT::v2i32) {
|
|
|
|
switch (Op.getOpcode()) {
|
|
|
|
case ISD::SRA:
|
2017-07-11 04:16:44 +08:00
|
|
|
Result = DAG.getNode(HexagonISD::VASR, dl, VT, V3, CommonSplat);
|
2015-03-20 00:33:08 +08:00
|
|
|
break;
|
|
|
|
case ISD::SHL:
|
2017-07-11 04:16:44 +08:00
|
|
|
Result = DAG.getNode(HexagonISD::VASL, dl, VT, V3, CommonSplat);
|
2015-03-20 00:33:08 +08:00
|
|
|
break;
|
|
|
|
case ISD::SRL:
|
2017-07-11 04:16:44 +08:00
|
|
|
Result = DAG.getNode(HexagonISD::VLSR, dl, VT, V3, CommonSplat);
|
2015-03-20 00:33:08 +08:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return SDValue();
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
return SDValue();
|
|
|
|
}
|
|
|
|
|
|
|
|
return DAG.getNode(ISD::BITCAST, dl, VT, Result);
|
|
|
|
}
|
|
|
|
|
|
|
|
SDValue
|
2017-11-23 04:56:23 +08:00
|
|
|
HexagonTargetLowering::buildVector32(ArrayRef<SDValue> Elem, const SDLoc &dl,
|
|
|
|
MVT VecTy, SelectionDAG &DAG) const {
|
|
|
|
MVT ElemTy = VecTy.getVectorElementType();
|
|
|
|
assert(VecTy.getVectorNumElements() == Elem.size());
|
|
|
|
|
|
|
|
SmallVector<ConstantSDNode*,4> Consts;
|
|
|
|
bool AllConst = true;
|
|
|
|
for (SDValue V : Elem) {
|
2017-12-07 00:40:37 +08:00
|
|
|
if (isUndef(V))
|
2017-11-23 04:56:23 +08:00
|
|
|
V = DAG.getConstant(0, dl, ElemTy);
|
|
|
|
auto *C = dyn_cast<ConstantSDNode>(V.getNode());
|
|
|
|
Consts.push_back(C);
|
|
|
|
AllConst = AllConst && C != nullptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned First, Num = Elem.size();
|
|
|
|
for (First = 0; First != Num; ++First)
|
2017-12-07 00:40:37 +08:00
|
|
|
if (!isUndef(Elem[First]))
|
2017-11-23 04:56:23 +08:00
|
|
|
break;
|
|
|
|
if (First == Num)
|
|
|
|
return DAG.getUNDEF(VecTy);
|
|
|
|
|
|
|
|
if (ElemTy == MVT::i16) {
|
|
|
|
assert(Elem.size() == 2);
|
|
|
|
if (AllConst) {
|
|
|
|
uint32_t V = (Consts[0]->getZExtValue() & 0xFFFF) |
|
|
|
|
Consts[1]->getZExtValue() << 16;
|
|
|
|
return DAG.getBitcast(MVT::v2i16, DAG.getConstant(V, dl, MVT::i32));
|
2017-07-14 02:17:58 +08:00
|
|
|
}
|
2017-12-07 00:40:37 +08:00
|
|
|
SDValue N = getNode(Hexagon::A2_combine_ll, dl, MVT::i32,
|
|
|
|
{Elem[1], Elem[0]}, DAG);
|
|
|
|
return DAG.getBitcast(MVT::v2i16, N);
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
|
|
|
|
2017-11-23 04:56:23 +08:00
|
|
|
// First try generating a constant.
|
|
|
|
assert(ElemTy == MVT::i8 && Num == 4);
|
|
|
|
if (AllConst) {
|
|
|
|
int32_t V = (Consts[0]->getZExtValue() & 0xFF) |
|
|
|
|
(Consts[1]->getZExtValue() & 0xFF) << 8 |
|
|
|
|
(Consts[1]->getZExtValue() & 0xFF) << 16 |
|
|
|
|
Consts[2]->getZExtValue() << 24;
|
|
|
|
return DAG.getBitcast(MVT::v4i8, DAG.getConstant(V, dl, MVT::i32));
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
|
|
|
|
2017-11-23 04:56:23 +08:00
|
|
|
// Then try splat.
|
|
|
|
bool IsSplat = true;
|
|
|
|
for (unsigned i = 0; i != Num; ++i) {
|
|
|
|
if (i == First)
|
2015-03-20 00:33:08 +08:00
|
|
|
continue;
|
2017-12-07 00:40:37 +08:00
|
|
|
if (Elem[i] == Elem[First] || isUndef(Elem[i]))
|
2017-11-23 04:56:23 +08:00
|
|
|
continue;
|
|
|
|
IsSplat = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (IsSplat)
|
|
|
|
return DAG.getNode(HexagonISD::VSPLAT, dl, VecTy, Elem[First]);
|
|
|
|
|
|
|
|
// Generate
|
|
|
|
// (zxtb(Elem[0]) | (zxtb(Elem[1]) << 8)) |
|
|
|
|
// (zxtb(Elem[2]) | (zxtb(Elem[3]) << 8)) << 16
|
|
|
|
SDValue S8 = DAG.getConstant(8, dl, MVT::i32);
|
2017-11-29 03:13:17 +08:00
|
|
|
SDValue V0 = DAG.getZeroExtendInReg(Elem[0], dl, MVT::i8);
|
|
|
|
SDValue V1 = DAG.getZeroExtendInReg(Elem[1], dl, MVT::i8);
|
|
|
|
SDValue V2 = DAG.getZeroExtendInReg(Elem[2], dl, MVT::i8);
|
|
|
|
SDValue V3 = DAG.getZeroExtendInReg(Elem[3], dl, MVT::i8);
|
|
|
|
|
|
|
|
SDValue V4 = DAG.getNode(ISD::SHL, dl, MVT::i32, {V1, S8});
|
|
|
|
SDValue V5 = DAG.getNode(ISD::SHL, dl, MVT::i32, {V3, S8});
|
|
|
|
SDValue V6 = DAG.getNode(ISD::OR, dl, MVT::i32, {V0, V4});
|
|
|
|
SDValue V7 = DAG.getNode(ISD::OR, dl, MVT::i32, {V2, V5});
|
2017-12-07 00:40:37 +08:00
|
|
|
|
|
|
|
SDValue T0 = getNode(Hexagon::A2_combine_ll, dl, MVT::i32, {V7, V6}, DAG);
|
|
|
|
return DAG.getBitcast(MVT::v4i8, T0);
|
2017-11-23 04:56:23 +08:00
|
|
|
}
|
2015-03-20 00:33:08 +08:00
|
|
|
|
2017-11-23 04:56:23 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::buildVector64(ArrayRef<SDValue> Elem, const SDLoc &dl,
|
|
|
|
MVT VecTy, SelectionDAG &DAG) const {
|
|
|
|
MVT ElemTy = VecTy.getVectorElementType();
|
|
|
|
assert(VecTy.getVectorNumElements() == Elem.size());
|
|
|
|
|
|
|
|
SmallVector<ConstantSDNode*,8> Consts;
|
|
|
|
bool AllConst = true;
|
|
|
|
for (SDValue V : Elem) {
|
2017-12-07 00:40:37 +08:00
|
|
|
if (isUndef(V))
|
2017-11-23 04:56:23 +08:00
|
|
|
V = DAG.getConstant(0, dl, ElemTy);
|
|
|
|
auto *C = dyn_cast<ConstantSDNode>(V.getNode());
|
|
|
|
Consts.push_back(C);
|
|
|
|
AllConst = AllConst && C != nullptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned First, Num = Elem.size();
|
|
|
|
for (First = 0; First != Num; ++First)
|
2017-12-07 00:40:37 +08:00
|
|
|
if (!isUndef(Elem[First]))
|
2017-11-23 04:56:23 +08:00
|
|
|
break;
|
|
|
|
if (First == Num)
|
|
|
|
return DAG.getUNDEF(VecTy);
|
|
|
|
|
|
|
|
// First try splat if possible.
|
|
|
|
if (ElemTy == MVT::i16) {
|
|
|
|
bool IsSplat = true;
|
|
|
|
for (unsigned i = 0; i != Num; ++i) {
|
|
|
|
if (i == First)
|
2015-03-20 00:33:08 +08:00
|
|
|
continue;
|
2017-12-07 00:40:37 +08:00
|
|
|
if (Elem[i] == Elem[First] || isUndef(Elem[i]))
|
2017-11-23 04:56:23 +08:00
|
|
|
continue;
|
|
|
|
IsSplat = false;
|
|
|
|
break;
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
2017-11-23 04:56:23 +08:00
|
|
|
if (IsSplat)
|
|
|
|
return DAG.getNode(HexagonISD::VSPLAT, dl, VecTy, Elem[First]);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Then try constant.
|
|
|
|
if (AllConst) {
|
|
|
|
uint64_t Val = 0;
|
|
|
|
unsigned W = ElemTy.getSizeInBits();
|
|
|
|
uint64_t Mask = (ElemTy == MVT::i8) ? 0xFFull
|
|
|
|
: (ElemTy == MVT::i16) ? 0xFFFFull : 0xFFFFFFFFull;
|
|
|
|
for (unsigned i = 0; i != Num; ++i)
|
|
|
|
Val = (Val << W) | (Consts[i]->getZExtValue() & Mask);
|
|
|
|
SDValue V0 = DAG.getConstant(Val, dl, MVT::i64);
|
|
|
|
return DAG.getBitcast(VecTy, V0);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Build two 32-bit vectors and concatenate.
|
|
|
|
MVT HalfTy = MVT::getVectorVT(ElemTy, Num/2);
|
|
|
|
SDValue L = (ElemTy == MVT::i32)
|
|
|
|
? Elem[0]
|
|
|
|
: buildVector32({Elem.data(), Num/2}, dl, HalfTy, DAG);
|
|
|
|
SDValue H = (ElemTy == MVT::i32)
|
|
|
|
? Elem[1]
|
|
|
|
: buildVector32({Elem.data()+Num/2, Num/2}, dl, HalfTy, DAG);
|
2017-12-07 00:40:37 +08:00
|
|
|
return DAG.getNode(HexagonISD::COMBINE, dl, VecTy, {H, L});
|
2017-11-23 04:56:23 +08:00
|
|
|
}
|
|
|
|
|
2017-11-30 03:58:10 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::extractVector(SDValue VecV, SDValue IdxV,
|
|
|
|
const SDLoc &dl, MVT ValTy, MVT ResTy,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
MVT VecTy = ty(VecV);
|
|
|
|
assert(!ValTy.isVector() ||
|
|
|
|
VecTy.getVectorElementType() == ValTy.getVectorElementType());
|
|
|
|
unsigned VecWidth = VecTy.getSizeInBits();
|
|
|
|
unsigned ValWidth = ValTy.getSizeInBits();
|
|
|
|
unsigned ElemWidth = VecTy.getVectorElementType().getSizeInBits();
|
|
|
|
assert(VecWidth == 32 || VecWidth == 64);
|
|
|
|
assert((VecWidth % ElemWidth) == 0);
|
|
|
|
|
|
|
|
// Cast everything to scalar integer types.
|
|
|
|
MVT ScalarTy = tyScalar(VecTy);
|
|
|
|
VecV = DAG.getBitcast(ScalarTy, VecV);
|
|
|
|
|
|
|
|
SDValue WidthV = DAG.getConstant(ValWidth, dl, MVT::i32);
|
|
|
|
SDValue ExtV;
|
|
|
|
|
|
|
|
if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(IdxV)) {
|
|
|
|
unsigned Off = C->getZExtValue() * ElemWidth;
|
|
|
|
if (VecWidth == 64 && ValWidth == 32) {
|
|
|
|
assert(Off == 0 || Off == 32);
|
|
|
|
unsigned SubIdx = Off == 0 ? Hexagon::isub_lo : Hexagon::isub_hi;
|
|
|
|
ExtV = DAG.getTargetExtractSubreg(SubIdx, dl, MVT::i32, VecV);
|
|
|
|
} else if (Off == 0 && (ValWidth % 8) == 0) {
|
|
|
|
ExtV = DAG.getZeroExtendInReg(VecV, dl, tyScalar(ValTy));
|
|
|
|
} else {
|
|
|
|
SDValue OffV = DAG.getConstant(Off, dl, MVT::i32);
|
|
|
|
// The return type of EXTRACTU must be the same as the type of the
|
|
|
|
// input vector.
|
|
|
|
ExtV = DAG.getNode(HexagonISD::EXTRACTU, dl, ScalarTy,
|
|
|
|
{VecV, WidthV, OffV});
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (ty(IdxV) != MVT::i32)
|
|
|
|
IdxV = DAG.getZExtOrTrunc(IdxV, dl, MVT::i32);
|
|
|
|
SDValue OffV = DAG.getNode(ISD::MUL, dl, MVT::i32, IdxV,
|
|
|
|
DAG.getConstant(ElemWidth, dl, MVT::i32));
|
|
|
|
// EXTRACTURP takes width/offset in a 64-bit pair.
|
|
|
|
SDValue CombV = DAG.getNode(HexagonISD::COMBINE, dl, MVT::i64,
|
|
|
|
{WidthV, OffV});
|
|
|
|
ExtV = DAG.getNode(HexagonISD::EXTRACTURP, dl, ScalarTy,
|
|
|
|
{VecV, CombV});
|
|
|
|
}
|
|
|
|
|
|
|
|
// Cast ExtV to the requested result type.
|
|
|
|
ExtV = DAG.getZExtOrTrunc(ExtV, dl, tyScalar(ResTy));
|
|
|
|
ExtV = DAG.getBitcast(ResTy, ExtV);
|
|
|
|
return ExtV;
|
|
|
|
}
|
|
|
|
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::insertVector(SDValue VecV, SDValue ValV, SDValue IdxV,
|
|
|
|
const SDLoc &dl, MVT ValTy,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
MVT VecTy = ty(VecV);
|
|
|
|
unsigned VecWidth = VecTy.getSizeInBits();
|
|
|
|
unsigned ValWidth = ValTy.getSizeInBits();
|
|
|
|
assert(VecWidth == 32 || VecWidth == 64);
|
|
|
|
assert((VecWidth % ValWidth) == 0);
|
|
|
|
|
|
|
|
// Cast everything to scalar integer types.
|
|
|
|
MVT ScalarTy = MVT::getIntegerVT(VecWidth);
|
|
|
|
// The actual type of ValV may be different than ValTy (which is related
|
|
|
|
// to the vector type).
|
|
|
|
unsigned VW = ty(ValV).getSizeInBits();
|
|
|
|
ValV = DAG.getBitcast(MVT::getIntegerVT(VW), ValV);
|
|
|
|
VecV = DAG.getBitcast(ScalarTy, VecV);
|
|
|
|
if (VW != VecWidth)
|
|
|
|
ValV = DAG.getAnyExtOrTrunc(ValV, dl, ScalarTy);
|
|
|
|
|
|
|
|
SDValue WidthV = DAG.getConstant(ValWidth, dl, MVT::i32);
|
|
|
|
SDValue InsV;
|
|
|
|
|
|
|
|
if (ConstantSDNode *C = dyn_cast<ConstantSDNode>(IdxV)) {
|
|
|
|
unsigned W = C->getZExtValue() * ValWidth;
|
|
|
|
SDValue OffV = DAG.getConstant(W, dl, MVT::i32);
|
|
|
|
InsV = DAG.getNode(HexagonISD::INSERT, dl, ScalarTy,
|
|
|
|
{VecV, ValV, WidthV, OffV});
|
|
|
|
} else {
|
|
|
|
if (ty(IdxV) != MVT::i32)
|
|
|
|
IdxV = DAG.getZExtOrTrunc(IdxV, dl, MVT::i32);
|
|
|
|
SDValue OffV = DAG.getNode(ISD::MUL, dl, MVT::i32, IdxV, WidthV);
|
|
|
|
// INSERTRP takes width/offset in a 64-bit pair.
|
|
|
|
SDValue CombV = DAG.getNode(HexagonISD::COMBINE, dl, MVT::i64,
|
|
|
|
{WidthV, OffV});
|
|
|
|
InsV = DAG.getNode(HexagonISD::INSERTRP, dl, ScalarTy,
|
|
|
|
{VecV, ValV, CombV});
|
|
|
|
}
|
|
|
|
|
|
|
|
return DAG.getNode(ISD::BITCAST, dl, VecTy, InsV);
|
|
|
|
}
|
|
|
|
|
2017-11-23 04:56:23 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerBUILD_VECTOR(SDValue Op, SelectionDAG &DAG) const {
|
2017-12-07 00:40:37 +08:00
|
|
|
MVT VecTy = ty(Op);
|
|
|
|
unsigned BW = VecTy.getSizeInBits();
|
2017-11-23 04:56:23 +08:00
|
|
|
if (BW == 32 || BW == 64) {
|
|
|
|
SmallVector<SDValue,8> Ops;
|
|
|
|
for (unsigned i = 0, e = Op.getNumOperands(); i != e; ++i)
|
|
|
|
Ops.push_back(Op.getOperand(i));
|
|
|
|
if (BW == 32)
|
2017-12-07 00:40:37 +08:00
|
|
|
return buildVector32(Ops, SDLoc(Op), VecTy, DAG);
|
|
|
|
return buildVector64(Ops, SDLoc(Op), VecTy, DAG);
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
|
|
|
|
2017-12-07 00:40:37 +08:00
|
|
|
if (Subtarget.useHVXOps() && Subtarget.isHVXVectorType(VecTy))
|
|
|
|
return LowerHvxBuildVector(Op, DAG);
|
|
|
|
|
2017-11-23 04:56:23 +08:00
|
|
|
return SDValue();
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerCONCAT_VECTORS(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
2017-12-07 00:40:37 +08:00
|
|
|
MVT VecTy = ty(Op);
|
|
|
|
assert(!Subtarget.useHVXOps() || !Subtarget.isHVXVectorType(VecTy));
|
2016-07-30 00:44:27 +08:00
|
|
|
|
2017-12-07 00:40:37 +08:00
|
|
|
if (VecTy.getSizeInBits() == 64) {
|
|
|
|
assert(Op.getNumOperands() == 2);
|
|
|
|
return DAG.getNode(HexagonISD::COMBINE, SDLoc(Op), VecTy, Op.getOperand(1),
|
|
|
|
Op.getOperand(0));
|
2016-07-30 00:44:27 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return SDValue();
|
|
|
|
}
|
|
|
|
|
2015-03-20 00:33:08 +08:00
|
|
|
SDValue
|
2017-11-30 03:58:10 +08:00
|
|
|
HexagonTargetLowering::LowerEXTRACT_VECTOR_ELT(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
2015-03-20 00:33:08 +08:00
|
|
|
SDValue Vec = Op.getOperand(0);
|
2017-12-07 00:40:37 +08:00
|
|
|
MVT VecTy = ty(Vec);
|
|
|
|
if (Subtarget.useHVXOps() && Subtarget.isHVXVectorType(VecTy))
|
|
|
|
return LowerHvxExtractElement(Op, DAG);
|
|
|
|
|
2017-11-30 03:58:10 +08:00
|
|
|
MVT ElemTy = ty(Vec).getVectorElementType();
|
|
|
|
return extractVector(Vec, Op.getOperand(1), SDLoc(Op), ElemTy, ty(Op), DAG);
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
SDValue
|
2017-11-30 03:58:10 +08:00
|
|
|
HexagonTargetLowering::LowerEXTRACT_SUBVECTOR(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
2015-03-20 00:33:08 +08:00
|
|
|
SDValue Vec = Op.getOperand(0);
|
2017-11-30 03:58:10 +08:00
|
|
|
MVT VecTy = ty(Vec);
|
|
|
|
if (Subtarget.useHVXOps() && Subtarget.isHVXVectorType(VecTy))
|
2017-12-07 00:40:37 +08:00
|
|
|
return LowerHvxExtractSubvector(Op, DAG);
|
2015-03-20 00:33:08 +08:00
|
|
|
|
2017-11-30 03:58:10 +08:00
|
|
|
return extractVector(Vec, Op.getOperand(1), SDLoc(Op), ty(Op), ty(Op), DAG);
|
|
|
|
}
|
2015-03-20 00:33:08 +08:00
|
|
|
|
2017-11-30 03:58:10 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerINSERT_VECTOR_ELT(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
MVT VecTy = ty(Op);
|
2017-12-07 00:40:37 +08:00
|
|
|
if (Subtarget.useHVXOps() && Subtarget.isHVXVectorType(VecTy))
|
|
|
|
return LowerHvxInsertElement(Op, DAG);
|
|
|
|
|
2017-11-30 03:58:10 +08:00
|
|
|
return insertVector(Op.getOperand(0), Op.getOperand(1), Op.getOperand(2),
|
|
|
|
SDLoc(Op), VecTy.getVectorElementType(), DAG);
|
|
|
|
}
|
2015-03-20 00:33:08 +08:00
|
|
|
|
2017-11-30 03:58:10 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerINSERT_SUBVECTOR(SDValue Op,
|
|
|
|
SelectionDAG &DAG) const {
|
2017-12-07 00:40:37 +08:00
|
|
|
if (Subtarget.useHVXOps() && Subtarget.isHVXVectorType(ty(Op)))
|
|
|
|
return LowerHvxInsertSubvector(Op, DAG);
|
|
|
|
|
2017-11-30 03:58:10 +08:00
|
|
|
SDValue ValV = Op.getOperand(1);
|
|
|
|
return insertVector(Op.getOperand(0), ValV, Op.getOperand(2),
|
|
|
|
SDLoc(Op), ty(ValV), DAG);
|
2015-03-20 00:33:08 +08:00
|
|
|
}
|
|
|
|
|
Refactor isInTailCallPosition handling
This change came about primarily because of two issues in the existing code.
Niether of:
define i64 @test1(i64 %val) {
%in = trunc i64 %val to i32
tail call i32 @ret32(i32 returned %in)
ret i64 %val
}
define i64 @test2(i64 %val) {
tail call i32 @ret32(i32 returned undef)
ret i32 42
}
should be tail calls, and the function sameNoopInput is responsible. The main
problem is that it is completely symmetric in the "tail call" and "ret" value,
but in reality different things are allowed on each side.
For these cases:
1. Any truncation should lead to a larger value being generated by "tail call"
than needed by "ret".
2. Undef should only be allowed as a source for ret, not as a result of the
call.
Along the way I noticed that a mismatch between what this function treats as a
valid truncation and what the backends see can lead to invalid calls as well
(see x86-32 test case).
This patch refactors the code so that instead of being based primarily on
values which it recurses into when necessary, it starts by inspecting the type
and considers each fundamental slot that the backend will see in turn. For
example, given a pathological function that returned {{}, {{}, i32, {}}, i32}
we would consider each "real" i32 in turn, and ask if it passes through
unchanged. This is much closer to what the backend sees as a result of
ComputeValueVTs.
Aside from the bug fixes, this eliminates the recursion that's going on and, I
believe, makes the bulk of the code significantly easier to understand. The
trade-off is the nasty iterators needed to find the real types inside a
returned value.
llvm-svn: 187787
2013-08-06 17:12:35 +08:00
|
|
|
bool
|
|
|
|
HexagonTargetLowering::allowTruncateForTailCall(Type *Ty1, Type *Ty2) const {
|
|
|
|
// Assuming the caller does not have either a signext or zeroext modifier, and
|
|
|
|
// only one value is accepted, any reasonable truncation is allowed.
|
|
|
|
if (!Ty1->isIntegerTy() || !Ty2->isIntegerTy())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// FIXME: in principle up to 64-bit could be made safe, but it would be very
|
|
|
|
// fragile at the moment: any support for multiple value returns would be
|
|
|
|
// liable to disallow tail calls involving i64 -> iN truncation in many cases.
|
|
|
|
return Ty1->getPrimitiveSizeInBits() <= 32;
|
|
|
|
}
|
|
|
|
|
2013-05-02 05:37:34 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerEH_RETURN(SDValue Op, SelectionDAG &DAG) const {
|
|
|
|
SDValue Chain = Op.getOperand(0);
|
|
|
|
SDValue Offset = Op.getOperand(1);
|
|
|
|
SDValue Handler = Op.getOperand(2);
|
2013-05-25 10:42:55 +08:00
|
|
|
SDLoc dl(Op);
|
2015-07-09 10:09:04 +08:00
|
|
|
auto PtrVT = getPointerTy(DAG.getDataLayout());
|
2013-05-02 05:37:34 +08:00
|
|
|
|
|
|
|
// Mark function as containing a call to EH_RETURN.
|
|
|
|
HexagonMachineFunctionInfo *FuncInfo =
|
|
|
|
DAG.getMachineFunction().getInfo<HexagonMachineFunctionInfo>();
|
|
|
|
FuncInfo->setHasEHReturn();
|
|
|
|
|
|
|
|
unsigned OffsetReg = Hexagon::R28;
|
|
|
|
|
2015-07-09 10:09:04 +08:00
|
|
|
SDValue StoreAddr =
|
|
|
|
DAG.getNode(ISD::ADD, dl, PtrVT, DAG.getRegister(Hexagon::R30, PtrVT),
|
|
|
|
DAG.getIntPtrConstant(4, dl));
|
[SelectionDAG] Get rid of bool parameters in SelectionDAG::getLoad, getStore, and friends.
Summary:
Instead, we take a single flags arg (a bitset).
Also add a default 0 alignment, and change the order of arguments so the
alignment comes before the flags.
This greatly simplifies many callsites, and fixes a bug in
AMDGPUISelLowering, wherein the order of the args to getLoad was
inverted. It also greatly simplifies the process of adding another flag
to getLoad.
Reviewers: chandlerc, tstellarAMD
Subscribers: jholewinski, arsenm, jyknight, dsanders, nemanjai, llvm-commits
Differential Revision: http://reviews.llvm.org/D22249
llvm-svn: 275592
2016-07-16 02:27:10 +08:00
|
|
|
Chain = DAG.getStore(Chain, dl, Handler, StoreAddr, MachinePointerInfo());
|
2013-05-02 05:37:34 +08:00
|
|
|
Chain = DAG.getCopyToReg(Chain, dl, OffsetReg, Offset);
|
|
|
|
|
|
|
|
// Not needed we already use it as explict input to EH_RETURN.
|
|
|
|
// MF.getRegInfo().addLiveOut(OffsetReg);
|
|
|
|
|
|
|
|
return DAG.getNode(HexagonISD::EH_RETURN, dl, MVT::Other, Chain);
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::LowerOperation(SDValue Op, SelectionDAG &DAG) const {
|
2015-04-23 05:17:00 +08:00
|
|
|
unsigned Opc = Op.getOpcode();
|
|
|
|
switch (Opc) {
|
|
|
|
default:
|
|
|
|
#ifndef NDEBUG
|
|
|
|
Op.getNode()->dumpr(&DAG);
|
|
|
|
if (Opc > HexagonISD::OP_BEGIN && Opc < HexagonISD::OP_END)
|
|
|
|
errs() << "Check for a non-legal type in this operation\n";
|
|
|
|
#endif
|
|
|
|
llvm_unreachable("Should not custom lower this!");
|
|
|
|
case ISD::CONCAT_VECTORS: return LowerCONCAT_VECTORS(Op, DAG);
|
2017-11-30 03:58:10 +08:00
|
|
|
case ISD::INSERT_SUBVECTOR: return LowerINSERT_SUBVECTOR(Op, DAG);
|
|
|
|
case ISD::INSERT_VECTOR_ELT: return LowerINSERT_VECTOR_ELT(Op, DAG);
|
|
|
|
case ISD::EXTRACT_SUBVECTOR: return LowerEXTRACT_SUBVECTOR(Op, DAG);
|
|
|
|
case ISD::EXTRACT_VECTOR_ELT: return LowerEXTRACT_VECTOR_ELT(Op, DAG);
|
2015-04-23 05:17:00 +08:00
|
|
|
case ISD::BUILD_VECTOR: return LowerBUILD_VECTOR(Op, DAG);
|
|
|
|
case ISD::VECTOR_SHUFFLE: return LowerVECTOR_SHUFFLE(Op, DAG);
|
2015-03-20 00:33:08 +08:00
|
|
|
case ISD::SRA:
|
|
|
|
case ISD::SHL:
|
2015-04-23 05:17:00 +08:00
|
|
|
case ISD::SRL: return LowerVECTOR_SHIFT(Op, DAG);
|
|
|
|
case ISD::ConstantPool: return LowerConstantPool(Op, DAG);
|
2015-12-19 04:19:30 +08:00
|
|
|
case ISD::JumpTable: return LowerJumpTable(Op, DAG);
|
2015-04-23 05:17:00 +08:00
|
|
|
case ISD::EH_RETURN: return LowerEH_RETURN(Op, DAG);
|
|
|
|
// Frame & Return address. Currently unimplemented.
|
|
|
|
case ISD::RETURNADDR: return LowerRETURNADDR(Op, DAG);
|
|
|
|
case ISD::FRAMEADDR: return LowerFRAMEADDR(Op, DAG);
|
2016-02-18 23:42:57 +08:00
|
|
|
case ISD::GlobalTLSAddress: return LowerGlobalTLSAddress(Op, DAG);
|
2015-04-23 05:17:00 +08:00
|
|
|
case ISD::ATOMIC_FENCE: return LowerATOMIC_FENCE(Op, DAG);
|
|
|
|
case ISD::GlobalAddress: return LowerGLOBALADDRESS(Op, DAG);
|
|
|
|
case ISD::BlockAddress: return LowerBlockAddress(Op, DAG);
|
2015-12-19 04:19:30 +08:00
|
|
|
case ISD::GLOBAL_OFFSET_TABLE: return LowerGLOBAL_OFFSET_TABLE(Op, DAG);
|
2015-04-23 05:17:00 +08:00
|
|
|
case ISD::VASTART: return LowerVASTART(Op, DAG);
|
|
|
|
case ISD::DYNAMIC_STACKALLOC: return LowerDYNAMIC_STACKALLOC(Op, DAG);
|
|
|
|
case ISD::SETCC: return LowerSETCC(Op, DAG);
|
|
|
|
case ISD::VSELECT: return LowerVSELECT(Op, DAG);
|
|
|
|
case ISD::INTRINSIC_WO_CHAIN: return LowerINTRINSIC_WO_CHAIN(Op, DAG);
|
2016-02-18 21:58:38 +08:00
|
|
|
case ISD::INTRINSIC_VOID: return LowerINTRINSIC_VOID(Op, DAG);
|
2015-04-23 05:17:00 +08:00
|
|
|
case ISD::INLINEASM: return LowerINLINEASM(Op, DAG);
|
2016-02-18 21:58:38 +08:00
|
|
|
case ISD::PREFETCH: return LowerPREFETCH(Op, DAG);
|
2017-02-23 06:28:47 +08:00
|
|
|
case ISD::READCYCLECOUNTER: return LowerREADCYCLECOUNTER(Op, DAG);
|
2017-12-08 01:37:28 +08:00
|
|
|
case ISD::MUL:
|
|
|
|
if (Subtarget.useHVXOps())
|
|
|
|
return LowerHvxMul(Op, DAG);
|
|
|
|
break;
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
2017-12-07 00:40:37 +08:00
|
|
|
return SDValue();
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
/// Returns relocation base for the given PIC jumptable.
|
|
|
|
SDValue
|
|
|
|
HexagonTargetLowering::getPICJumpTableRelocBase(SDValue Table,
|
|
|
|
SelectionDAG &DAG) const {
|
|
|
|
int Idx = cast<JumpTableSDNode>(Table)->getIndex();
|
|
|
|
EVT VT = Table.getValueType();
|
|
|
|
SDValue T = DAG.getTargetJumpTable(Idx, VT, HexagonII::MO_PCREL);
|
|
|
|
return DAG.getNode(HexagonISD::AT_PCREL, SDLoc(Table), VT, T);
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
// Inline Assembly Support
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2016-05-18 22:34:51 +08:00
|
|
|
TargetLowering::ConstraintType
|
|
|
|
HexagonTargetLowering::getConstraintType(StringRef Constraint) const {
|
|
|
|
if (Constraint.size() == 1) {
|
|
|
|
switch (Constraint[0]) {
|
|
|
|
case 'q':
|
|
|
|
case 'v':
|
|
|
|
if (Subtarget.useHVXOps())
|
2017-07-22 01:51:27 +08:00
|
|
|
return C_RegisterClass;
|
|
|
|
break;
|
|
|
|
case 'a':
|
|
|
|
return C_RegisterClass;
|
|
|
|
default:
|
2016-05-18 22:34:51 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return TargetLowering::getConstraintType(Constraint);
|
|
|
|
}
|
|
|
|
|
2016-08-20 03:29:15 +08:00
|
|
|
std::pair<unsigned, const TargetRegisterClass*>
|
2015-02-27 06:38:43 +08:00
|
|
|
HexagonTargetLowering::getRegForInlineAsmConstraint(
|
2015-07-06 03:29:18 +08:00
|
|
|
const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const {
|
2015-11-27 02:38:27 +08:00
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
if (Constraint.size() == 1) {
|
|
|
|
switch (Constraint[0]) {
|
|
|
|
case 'r': // R0-R31
|
2016-08-20 03:29:15 +08:00
|
|
|
switch (VT.SimpleTy) {
|
|
|
|
default:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, nullptr};
|
2016-08-20 03:29:15 +08:00
|
|
|
case MVT::i1:
|
|
|
|
case MVT::i8:
|
|
|
|
case MVT::i16:
|
|
|
|
case MVT::i32:
|
|
|
|
case MVT::f32:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, &Hexagon::IntRegsRegClass};
|
2016-08-20 03:29:15 +08:00
|
|
|
case MVT::i64:
|
|
|
|
case MVT::f64:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, &Hexagon::DoubleRegsRegClass};
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
2017-07-22 01:51:27 +08:00
|
|
|
break;
|
|
|
|
case 'a': // M0-M1
|
2017-10-21 04:24:44 +08:00
|
|
|
if (VT != MVT::i32)
|
|
|
|
return {0u, nullptr};
|
|
|
|
return {0u, &Hexagon::ModRegsRegClass};
|
2015-11-27 02:38:27 +08:00
|
|
|
case 'q': // q0-q3
|
2017-03-03 01:50:24 +08:00
|
|
|
switch (VT.getSizeInBits()) {
|
2016-08-20 03:29:15 +08:00
|
|
|
default:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, nullptr};
|
2017-03-03 01:50:24 +08:00
|
|
|
case 512:
|
|
|
|
case 1024:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, &Hexagon::HvxQRRegClass};
|
2016-08-20 03:29:15 +08:00
|
|
|
}
|
2017-07-22 01:51:27 +08:00
|
|
|
break;
|
2015-11-27 02:38:27 +08:00
|
|
|
case 'v': // V0-V31
|
2017-03-03 01:50:24 +08:00
|
|
|
switch (VT.getSizeInBits()) {
|
2016-08-20 03:29:15 +08:00
|
|
|
default:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, nullptr};
|
2017-03-03 01:50:24 +08:00
|
|
|
case 512:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, &Hexagon::HvxVRRegClass};
|
2017-03-03 01:50:24 +08:00
|
|
|
case 1024:
|
2017-10-19 02:07:07 +08:00
|
|
|
if (Subtarget.hasV60TOps() && Subtarget.useHVX128BOps())
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, &Hexagon::HvxVRRegClass};
|
|
|
|
return {0u, &Hexagon::HvxWRRegClass};
|
2017-03-03 01:50:24 +08:00
|
|
|
case 2048:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, &Hexagon::HvxWRRegClass};
|
2016-08-20 03:29:15 +08:00
|
|
|
}
|
2017-07-22 01:51:27 +08:00
|
|
|
break;
|
2011-12-13 05:14:40 +08:00
|
|
|
default:
|
2017-10-21 04:24:44 +08:00
|
|
|
return {0u, nullptr};
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-02-27 06:38:43 +08:00
|
|
|
return TargetLowering::getRegForInlineAsmConstraint(TRI, Constraint, VT);
|
2011-12-13 05:14:40 +08:00
|
|
|
}
|
|
|
|
|
2012-05-11 04:20:25 +08:00
|
|
|
/// isFPImmLegal - Returns true if the target can instruction select the
|
|
|
|
/// specified FP immediate natively. If false, the legalizer will
|
|
|
|
/// materialize the FP immediate as a load from a constant pool.
|
|
|
|
bool HexagonTargetLowering::isFPImmLegal(const APFloat &Imm, EVT VT) const {
|
2015-04-23 05:17:00 +08:00
|
|
|
return Subtarget.hasV5TOps();
|
2012-05-11 04:20:25 +08:00
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
/// isLegalAddressingMode - Return true if the addressing mode represented by
|
|
|
|
/// AM is legal for this target, for a load/store of the specified type.
|
2015-07-09 10:09:40 +08:00
|
|
|
bool HexagonTargetLowering::isLegalAddressingMode(const DataLayout &DL,
|
|
|
|
const AddrMode &AM, Type *Ty,
|
2017-07-21 19:59:37 +08:00
|
|
|
unsigned AS, Instruction *I) const {
|
2016-08-03 23:06:18 +08:00
|
|
|
if (Ty->isSized()) {
|
|
|
|
// When LSR detects uses of the same base address to access different
|
|
|
|
// types (e.g. unions), it will assume a conservative type for these
|
|
|
|
// uses:
|
|
|
|
// LSR Use: Kind=Address of void in addrspace(4294967295), ...
|
|
|
|
// The type Ty passed here would then be "void". Skip the alignment
|
|
|
|
// checks, but do not return false right away, since that confuses
|
|
|
|
// LSR into crashing.
|
|
|
|
unsigned A = DL.getABITypeAlignment(Ty);
|
|
|
|
// The base offset must be a multiple of the alignment.
|
|
|
|
if ((AM.BaseOffs % A) != 0)
|
|
|
|
return false;
|
|
|
|
// The shifted offset must fit in 11 bits.
|
|
|
|
if (!isInt<11>(AM.BaseOffs >> Log2_32(A)))
|
|
|
|
return false;
|
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// No global is ever allowed as a base.
|
2015-04-23 05:17:00 +08:00
|
|
|
if (AM.BaseGV)
|
2011-12-13 05:14:40 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
int Scale = AM.Scale;
|
2016-08-02 04:31:50 +08:00
|
|
|
if (Scale < 0)
|
|
|
|
Scale = -Scale;
|
2011-12-13 05:14:40 +08:00
|
|
|
switch (Scale) {
|
|
|
|
case 0: // No scale reg, "r+i", "r", or just "i".
|
|
|
|
break;
|
|
|
|
default: // No scaled addressing mode.
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2015-12-19 04:19:30 +08:00
|
|
|
/// Return true if folding a constant offset with the given GlobalAddress is
|
|
|
|
/// legal. It is frequently not legal in PIC relocation models.
|
|
|
|
bool HexagonTargetLowering::isOffsetFoldingLegal(const GlobalAddressSDNode *GA)
|
|
|
|
const {
|
|
|
|
return HTM.getRelocationModel() == Reloc::Static;
|
|
|
|
}
|
|
|
|
|
2011-12-13 05:14:40 +08:00
|
|
|
/// isLegalICmpImmediate - Return true if the specified immediate is legal
|
|
|
|
/// icmp immediate, that is the target has icmp instructions which can compare
|
|
|
|
/// a register against the immediate without having to materialize the
|
|
|
|
/// immediate into a register.
|
|
|
|
bool HexagonTargetLowering::isLegalICmpImmediate(int64_t Imm) const {
|
|
|
|
return Imm >= -512 && Imm <= 511;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// IsEligibleForTailCallOptimization - Check whether the call is eligible
|
|
|
|
/// for tail call optimization. Targets which want to do tail call
|
|
|
|
/// optimization should implement this function.
|
|
|
|
bool HexagonTargetLowering::IsEligibleForTailCallOptimization(
|
|
|
|
SDValue Callee,
|
|
|
|
CallingConv::ID CalleeCC,
|
|
|
|
bool isVarArg,
|
|
|
|
bool isCalleeStructRet,
|
|
|
|
bool isCallerStructRet,
|
|
|
|
const SmallVectorImpl<ISD::OutputArg> &Outs,
|
|
|
|
const SmallVectorImpl<SDValue> &OutVals,
|
|
|
|
const SmallVectorImpl<ISD::InputArg> &Ins,
|
|
|
|
SelectionDAG& DAG) const {
|
2017-12-16 06:22:58 +08:00
|
|
|
const Function &CallerF = DAG.getMachineFunction().getFunction();
|
|
|
|
CallingConv::ID CallerCC = CallerF.getCallingConv();
|
2011-12-13 05:14:40 +08:00
|
|
|
bool CCMatch = CallerCC == CalleeCC;
|
|
|
|
|
|
|
|
// ***************************************************************************
|
|
|
|
// Look for obvious safe cases to perform tail call optimization that do not
|
|
|
|
// require ABI changes.
|
|
|
|
// ***************************************************************************
|
|
|
|
|
|
|
|
// If this is a tail call via a function pointer, then don't do it!
|
2016-08-02 04:31:50 +08:00
|
|
|
if (!isa<GlobalAddressSDNode>(Callee) &&
|
|
|
|
!isa<ExternalSymbolSDNode>(Callee)) {
|
2011-12-13 05:14:40 +08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2016-08-19 23:02:18 +08:00
|
|
|
// Do not optimize if the calling conventions do not match and the conventions
|
|
|
|
// used are not C or Fast.
|
|
|
|
if (!CCMatch) {
|
|
|
|
bool R = (CallerCC == CallingConv::C || CallerCC == CallingConv::Fast);
|
|
|
|
bool E = (CalleeCC == CallingConv::C || CalleeCC == CallingConv::Fast);
|
|
|
|
// If R & E, then ok.
|
|
|
|
if (!R || !E)
|
|
|
|
return false;
|
|
|
|
}
|
2011-12-13 05:14:40 +08:00
|
|
|
|
|
|
|
// Do not tail call optimize vararg calls.
|
|
|
|
if (isVarArg)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Also avoid tail call optimization if either caller or callee uses struct
|
|
|
|
// return semantics.
|
|
|
|
if (isCalleeStructRet || isCallerStructRet)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// In addition to the cases above, we also disable Tail Call Optimization if
|
|
|
|
// the calling convention code that at least one outgoing argument needs to
|
|
|
|
// go on the stack. We cannot check that here because at this point that
|
|
|
|
// information is not available.
|
|
|
|
return true;
|
|
|
|
}
|
2014-12-09 05:19:18 +08:00
|
|
|
|
2016-08-03 02:34:31 +08:00
|
|
|
/// Returns the target specific optimal type for load and store operations as
|
|
|
|
/// a result of memset, memcpy, and memmove lowering.
|
|
|
|
///
|
|
|
|
/// If DstAlign is zero that means it's safe to destination alignment can
|
|
|
|
/// satisfy any constraint. Similarly if SrcAlign is zero it means there isn't
|
|
|
|
/// a need to check it against alignment requirement, probably because the
|
|
|
|
/// source does not need to be loaded. If 'IsMemset' is true, that means it's
|
|
|
|
/// expanding a memset. If 'ZeroMemset' is true, that means it's a memset of
|
|
|
|
/// zero. 'MemcpyStrSrc' indicates whether the memcpy source is constant so it
|
|
|
|
/// does not need to be loaded. It returns EVT::Other if the type should be
|
|
|
|
/// determined using generic target-independent logic.
|
|
|
|
EVT HexagonTargetLowering::getOptimalMemOpType(uint64_t Size,
|
|
|
|
unsigned DstAlign, unsigned SrcAlign, bool IsMemset, bool ZeroMemset,
|
|
|
|
bool MemcpyStrSrc, MachineFunction &MF) const {
|
|
|
|
|
|
|
|
auto Aligned = [](unsigned GivenA, unsigned MinA) -> bool {
|
|
|
|
return (GivenA % MinA) == 0;
|
|
|
|
};
|
|
|
|
|
|
|
|
if (Size >= 8 && Aligned(DstAlign, 8) && (IsMemset || Aligned(SrcAlign, 8)))
|
|
|
|
return MVT::i64;
|
|
|
|
if (Size >= 4 && Aligned(DstAlign, 4) && (IsMemset || Aligned(SrcAlign, 4)))
|
|
|
|
return MVT::i32;
|
|
|
|
if (Size >= 2 && Aligned(DstAlign, 2) && (IsMemset || Aligned(SrcAlign, 2)))
|
|
|
|
return MVT::i16;
|
|
|
|
|
|
|
|
return MVT::Other;
|
|
|
|
}
|
|
|
|
|
2016-03-28 23:43:03 +08:00
|
|
|
bool HexagonTargetLowering::allowsMisalignedMemoryAccesses(EVT VT,
|
|
|
|
unsigned AS, unsigned Align, bool *Fast) const {
|
|
|
|
if (Fast)
|
|
|
|
*Fast = false;
|
|
|
|
|
|
|
|
switch (VT.getSimpleVT().SimpleTy) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case MVT::v64i8:
|
|
|
|
case MVT::v128i8:
|
|
|
|
case MVT::v256i8:
|
|
|
|
case MVT::v32i16:
|
|
|
|
case MVT::v64i16:
|
|
|
|
case MVT::v128i16:
|
|
|
|
case MVT::v16i32:
|
|
|
|
case MVT::v32i32:
|
|
|
|
case MVT::v64i32:
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2015-11-27 02:38:27 +08:00
|
|
|
std::pair<const TargetRegisterClass*, uint8_t>
|
|
|
|
HexagonTargetLowering::findRepresentativeClass(const TargetRegisterInfo *TRI,
|
|
|
|
MVT VT) const {
|
|
|
|
const TargetRegisterClass *RRC = nullptr;
|
|
|
|
|
|
|
|
uint8_t Cost = 1;
|
|
|
|
switch (VT.SimpleTy) {
|
|
|
|
default:
|
|
|
|
return TargetLowering::findRepresentativeClass(TRI, VT);
|
|
|
|
case MVT::v64i8:
|
|
|
|
case MVT::v32i16:
|
|
|
|
case MVT::v16i32:
|
2017-09-15 23:46:05 +08:00
|
|
|
RRC = &Hexagon::HvxVRRegClass;
|
2015-11-27 02:38:27 +08:00
|
|
|
break;
|
|
|
|
case MVT::v128i8:
|
|
|
|
case MVT::v64i16:
|
|
|
|
case MVT::v32i32:
|
|
|
|
if (Subtarget.hasV60TOps() && Subtarget.useHVXOps() &&
|
2017-10-19 02:07:07 +08:00
|
|
|
Subtarget.useHVX128BOps())
|
2017-09-15 23:46:05 +08:00
|
|
|
RRC = &Hexagon::HvxVRRegClass;
|
2015-11-27 02:38:27 +08:00
|
|
|
else
|
2017-09-15 23:46:05 +08:00
|
|
|
RRC = &Hexagon::HvxWRRegClass;
|
2015-11-27 02:38:27 +08:00
|
|
|
break;
|
|
|
|
case MVT::v256i8:
|
|
|
|
case MVT::v128i16:
|
|
|
|
case MVT::v64i32:
|
2017-09-15 23:46:05 +08:00
|
|
|
RRC = &Hexagon::HvxWRRegClass;
|
2015-11-27 02:38:27 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
return std::make_pair(RRC, Cost);
|
|
|
|
}
|
|
|
|
|
2015-07-09 22:51:21 +08:00
|
|
|
Value *HexagonTargetLowering::emitLoadLinked(IRBuilder<> &Builder, Value *Addr,
|
|
|
|
AtomicOrdering Ord) const {
|
|
|
|
BasicBlock *BB = Builder.GetInsertBlock();
|
|
|
|
Module *M = BB->getParent()->getParent();
|
|
|
|
Type *Ty = cast<PointerType>(Addr->getType())->getElementType();
|
|
|
|
unsigned SZ = Ty->getPrimitiveSizeInBits();
|
|
|
|
assert((SZ == 32 || SZ == 64) && "Only 32/64-bit atomic loads supported");
|
|
|
|
Intrinsic::ID IntID = (SZ == 32) ? Intrinsic::hexagon_L2_loadw_locked
|
|
|
|
: Intrinsic::hexagon_L4_loadd_locked;
|
|
|
|
Value *Fn = Intrinsic::getDeclaration(M, IntID);
|
|
|
|
return Builder.CreateCall(Fn, Addr, "larx");
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Perform a store-conditional operation to Addr. Return the status of the
|
|
|
|
/// store. This should be 0 if the store succeeded, non-zero otherwise.
|
|
|
|
Value *HexagonTargetLowering::emitStoreConditional(IRBuilder<> &Builder,
|
|
|
|
Value *Val, Value *Addr, AtomicOrdering Ord) const {
|
|
|
|
BasicBlock *BB = Builder.GetInsertBlock();
|
|
|
|
Module *M = BB->getParent()->getParent();
|
|
|
|
Type *Ty = Val->getType();
|
|
|
|
unsigned SZ = Ty->getPrimitiveSizeInBits();
|
|
|
|
assert((SZ == 32 || SZ == 64) && "Only 32/64-bit atomic stores supported");
|
|
|
|
Intrinsic::ID IntID = (SZ == 32) ? Intrinsic::hexagon_S2_storew_locked
|
|
|
|
: Intrinsic::hexagon_S4_stored_locked;
|
|
|
|
Value *Fn = Intrinsic::getDeclaration(M, IntID);
|
|
|
|
Value *Call = Builder.CreateCall(Fn, {Addr, Val}, "stcx");
|
|
|
|
Value *Cmp = Builder.CreateICmpEQ(Call, Builder.getInt32(0), "");
|
|
|
|
Value *Ext = Builder.CreateZExt(Cmp, Type::getInt32Ty(M->getContext()));
|
|
|
|
return Ext;
|
|
|
|
}
|
|
|
|
|
2015-09-12 01:08:28 +08:00
|
|
|
TargetLowering::AtomicExpansionKind
|
|
|
|
HexagonTargetLowering::shouldExpandAtomicLoadInIR(LoadInst *LI) const {
|
2015-07-09 22:51:21 +08:00
|
|
|
// Do not expand loads and stores that don't exceed 64 bits.
|
2015-09-12 01:08:28 +08:00
|
|
|
return LI->getType()->getPrimitiveSizeInBits() > 64
|
2015-12-03 02:12:57 +08:00
|
|
|
? AtomicExpansionKind::LLOnly
|
2015-09-12 01:08:28 +08:00
|
|
|
: AtomicExpansionKind::None;
|
2015-07-09 22:51:21 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bool HexagonTargetLowering::shouldExpandAtomicStoreInIR(StoreInst *SI) const {
|
|
|
|
// Do not expand loads and stores that don't exceed 64 bits.
|
|
|
|
return SI->getValueOperand()->getType()->getPrimitiveSizeInBits() > 64;
|
|
|
|
}
|
2016-06-23 00:07:10 +08:00
|
|
|
|
|
|
|
bool HexagonTargetLowering::shouldExpandAtomicCmpXchgInIR(
|
|
|
|
AtomicCmpXchgInst *AI) const {
|
|
|
|
const DataLayout &DL = AI->getModule()->getDataLayout();
|
|
|
|
unsigned Size = DL.getTypeStoreSize(AI->getCompareOperand()->getType());
|
|
|
|
return Size >= 4 && Size <= 8;
|
|
|
|
}
|