Change TargetLowering::canMergeStoresTo() to take a MF instead of DAG.

DAG is unnecessary and we need this hook to implement store merging on GlobalISel too.
This commit is contained in:
Amara Emerson 2021-08-06 12:42:03 -07:00
parent 71ae2e0221
commit 2b067e3335
10 changed files with 29 additions and 20 deletions

View File

@ -591,7 +591,7 @@ public:
/// Returns if it's reasonable to merge stores to MemVT size. /// Returns if it's reasonable to merge stores to MemVT size.
virtual bool canMergeStoresTo(unsigned AS, EVT MemVT, virtual bool canMergeStoresTo(unsigned AS, EVT MemVT,
const SelectionDAG &DAG) const { const MachineFunction &MF) const {
return true; return true;
} }

View File

@ -17394,7 +17394,8 @@ bool DAGCombiner::tryStoreMergeOfConstants(
break; break;
if (TLI.isTypeLegal(StoreTy) && if (TLI.isTypeLegal(StoreTy) &&
TLI.canMergeStoresTo(FirstStoreAS, StoreTy, DAG) && TLI.canMergeStoresTo(FirstStoreAS, StoreTy,
DAG.getMachineFunction()) &&
TLI.allowsMemoryAccess(Context, DL, StoreTy, TLI.allowsMemoryAccess(Context, DL, StoreTy,
*FirstInChain->getMemOperand(), &IsFast) && *FirstInChain->getMemOperand(), &IsFast) &&
IsFast) { IsFast) {
@ -17406,7 +17407,8 @@ bool DAGCombiner::tryStoreMergeOfConstants(
EVT LegalizedStoredValTy = EVT LegalizedStoredValTy =
TLI.getTypeToTransformTo(Context, StoredVal.getValueType()); TLI.getTypeToTransformTo(Context, StoredVal.getValueType());
if (TLI.isTruncStoreLegal(LegalizedStoredValTy, StoreTy) && if (TLI.isTruncStoreLegal(LegalizedStoredValTy, StoreTy) &&
TLI.canMergeStoresTo(FirstStoreAS, LegalizedStoredValTy, DAG) && TLI.canMergeStoresTo(FirstStoreAS, LegalizedStoredValTy,
DAG.getMachineFunction()) &&
TLI.allowsMemoryAccess(Context, DL, StoreTy, TLI.allowsMemoryAccess(Context, DL, StoreTy,
*FirstInChain->getMemOperand(), &IsFast) && *FirstInChain->getMemOperand(), &IsFast) &&
IsFast) { IsFast) {
@ -17425,7 +17427,7 @@ bool DAGCombiner::tryStoreMergeOfConstants(
unsigned Elts = (i + 1) * NumMemElts; unsigned Elts = (i + 1) * NumMemElts;
EVT Ty = EVT::getVectorVT(Context, MemVT.getScalarType(), Elts); EVT Ty = EVT::getVectorVT(Context, MemVT.getScalarType(), Elts);
if (TLI.isTypeLegal(Ty) && TLI.isTypeLegal(MemVT) && if (TLI.isTypeLegal(Ty) && TLI.isTypeLegal(MemVT) &&
TLI.canMergeStoresTo(FirstStoreAS, Ty, DAG) && TLI.canMergeStoresTo(FirstStoreAS, Ty, DAG.getMachineFunction()) &&
TLI.allowsMemoryAccess(Context, DL, Ty, TLI.allowsMemoryAccess(Context, DL, Ty,
*FirstInChain->getMemOperand(), &IsFast) && *FirstInChain->getMemOperand(), &IsFast) &&
IsFast) IsFast)
@ -17501,7 +17503,8 @@ bool DAGCombiner::tryStoreMergeOfExtracts(
if (Ty.getSizeInBits() > MaximumLegalStoreInBits) if (Ty.getSizeInBits() > MaximumLegalStoreInBits)
break; break;
if (TLI.isTypeLegal(Ty) && TLI.canMergeStoresTo(FirstStoreAS, Ty, DAG) && if (TLI.isTypeLegal(Ty) &&
TLI.canMergeStoresTo(FirstStoreAS, Ty, DAG.getMachineFunction()) &&
TLI.allowsMemoryAccess(Context, DL, Ty, TLI.allowsMemoryAccess(Context, DL, Ty,
*FirstInChain->getMemOperand(), &IsFast) && *FirstInChain->getMemOperand(), &IsFast) &&
IsFast) IsFast)
@ -17650,7 +17653,8 @@ bool DAGCombiner::tryStoreMergeOfLoads(SmallVectorImpl<MemOpLink> &StoreNodes,
bool IsFastSt = false; bool IsFastSt = false;
bool IsFastLd = false; bool IsFastLd = false;
if (TLI.isTypeLegal(StoreTy) && if (TLI.isTypeLegal(StoreTy) &&
TLI.canMergeStoresTo(FirstStoreAS, StoreTy, DAG) && TLI.canMergeStoresTo(FirstStoreAS, StoreTy,
DAG.getMachineFunction()) &&
TLI.allowsMemoryAccess(Context, DL, StoreTy, TLI.allowsMemoryAccess(Context, DL, StoreTy,
*FirstInChain->getMemOperand(), &IsFastSt) && *FirstInChain->getMemOperand(), &IsFastSt) &&
IsFastSt && IsFastSt &&
@ -17664,7 +17668,8 @@ bool DAGCombiner::tryStoreMergeOfLoads(SmallVectorImpl<MemOpLink> &StoreNodes,
unsigned SizeInBits = (i + 1) * ElementSizeBytes * 8; unsigned SizeInBits = (i + 1) * ElementSizeBytes * 8;
StoreTy = EVT::getIntegerVT(Context, SizeInBits); StoreTy = EVT::getIntegerVT(Context, SizeInBits);
if (TLI.isTypeLegal(StoreTy) && if (TLI.isTypeLegal(StoreTy) &&
TLI.canMergeStoresTo(FirstStoreAS, StoreTy, DAG) && TLI.canMergeStoresTo(FirstStoreAS, StoreTy,
DAG.getMachineFunction()) &&
TLI.allowsMemoryAccess(Context, DL, StoreTy, TLI.allowsMemoryAccess(Context, DL, StoreTy,
*FirstInChain->getMemOperand(), &IsFastSt) && *FirstInChain->getMemOperand(), &IsFastSt) &&
IsFastSt && IsFastSt &&
@ -17678,7 +17683,8 @@ bool DAGCombiner::tryStoreMergeOfLoads(SmallVectorImpl<MemOpLink> &StoreNodes,
TargetLowering::TypePromoteInteger) { TargetLowering::TypePromoteInteger) {
EVT LegalizedStoredValTy = TLI.getTypeToTransformTo(Context, StoreTy); EVT LegalizedStoredValTy = TLI.getTypeToTransformTo(Context, StoreTy);
if (TLI.isTruncStoreLegal(LegalizedStoredValTy, StoreTy) && if (TLI.isTruncStoreLegal(LegalizedStoredValTy, StoreTy) &&
TLI.canMergeStoresTo(FirstStoreAS, LegalizedStoredValTy, DAG) && TLI.canMergeStoresTo(FirstStoreAS, LegalizedStoredValTy,
DAG.getMachineFunction()) &&
TLI.isLoadExtLegal(ISD::ZEXTLOAD, LegalizedStoredValTy, StoreTy) && TLI.isLoadExtLegal(ISD::ZEXTLOAD, LegalizedStoredValTy, StoreTy) &&
TLI.isLoadExtLegal(ISD::SEXTLOAD, LegalizedStoredValTy, StoreTy) && TLI.isLoadExtLegal(ISD::SEXTLOAD, LegalizedStoredValTy, StoreTy) &&
TLI.isLoadExtLegal(ISD::EXTLOAD, LegalizedStoredValTy, StoreTy) && TLI.isLoadExtLegal(ISD::EXTLOAD, LegalizedStoredValTy, StoreTy) &&

View File

@ -701,12 +701,11 @@ public:
bool isIntDivCheap(EVT VT, AttributeList Attr) const override; bool isIntDivCheap(EVT VT, AttributeList Attr) const override;
bool canMergeStoresTo(unsigned AddressSpace, EVT MemVT, bool canMergeStoresTo(unsigned AddressSpace, EVT MemVT,
const SelectionDAG &DAG) const override { const MachineFunction &MF) const override {
// Do not merge to float value size (128 bytes) if no implicit // Do not merge to float value size (128 bytes) if no implicit
// float attribute is set. // float attribute is set.
bool NoFloat = DAG.getMachineFunction().getFunction().hasFnAttribute( bool NoFloat = MF.getFunction().hasFnAttribute(Attribute::NoImplicitFloat);
Attribute::NoImplicitFloat);
if (NoFloat) if (NoFloat)
return (MemVT.getSizeInBits() <= 64); return (MemVT.getSizeInBits() <= 64);

View File

@ -18,6 +18,7 @@
#include "R600InstrInfo.h" #include "R600InstrInfo.h"
#include "R600MachineFunctionInfo.h" #include "R600MachineFunctionInfo.h"
#include "R600Subtarget.h" #include "R600Subtarget.h"
#include "llvm/CodeGen/MachineFunction.h"
#include "llvm/IR/IntrinsicsAMDGPU.h" #include "llvm/IR/IntrinsicsAMDGPU.h"
#include "llvm/IR/IntrinsicsR600.h" #include "llvm/IR/IntrinsicsR600.h"
@ -1564,7 +1565,7 @@ EVT R600TargetLowering::getSetCCResultType(const DataLayout &DL, LLVMContext &,
} }
bool R600TargetLowering::canMergeStoresTo(unsigned AS, EVT MemVT, bool R600TargetLowering::canMergeStoresTo(unsigned AS, EVT MemVT,
const SelectionDAG &DAG) const { const MachineFunction &MF) const {
// Local and Private addresses do not handle vectors. Limit to i32 // Local and Private addresses do not handle vectors. Limit to i32
if ((AS == AMDGPUAS::LOCAL_ADDRESS || AS == AMDGPUAS::PRIVATE_ADDRESS)) { if ((AS == AMDGPUAS::LOCAL_ADDRESS || AS == AMDGPUAS::PRIVATE_ADDRESS)) {
return (MemVT.getSizeInBits() <= 32); return (MemVT.getSizeInBits() <= 32);

View File

@ -15,6 +15,7 @@
#define LLVM_LIB_TARGET_AMDGPU_R600ISELLOWERING_H #define LLVM_LIB_TARGET_AMDGPU_R600ISELLOWERING_H
#include "AMDGPUISelLowering.h" #include "AMDGPUISelLowering.h"
#include "llvm/CodeGen/MachineFunction.h"
namespace llvm { namespace llvm {
@ -47,7 +48,7 @@ public:
EVT VT) const override; EVT VT) const override;
bool canMergeStoresTo(unsigned AS, EVT MemVT, bool canMergeStoresTo(unsigned AS, EVT MemVT,
const SelectionDAG &DAG) const override; const MachineFunction &MF) const override;
bool allowsMisalignedMemoryAccesses( bool allowsMisalignedMemoryAccesses(
EVT VT, unsigned AS, Align Alignment, EVT VT, unsigned AS, Align Alignment,

View File

@ -23,6 +23,7 @@
#include "llvm/CodeGen/Analysis.h" #include "llvm/CodeGen/Analysis.h"
#include "llvm/CodeGen/FunctionLoweringInfo.h" #include "llvm/CodeGen/FunctionLoweringInfo.h"
#include "llvm/CodeGen/GlobalISel/GISelKnownBits.h" #include "llvm/CodeGen/GlobalISel/GISelKnownBits.h"
#include "llvm/CodeGen/MachineFunction.h"
#include "llvm/CodeGen/MachineLoopInfo.h" #include "llvm/CodeGen/MachineLoopInfo.h"
#include "llvm/IR/DiagnosticInfo.h" #include "llvm/IR/DiagnosticInfo.h"
#include "llvm/IR/IntrinsicInst.h" #include "llvm/IR/IntrinsicInst.h"
@ -1427,7 +1428,7 @@ bool SITargetLowering::isLegalAddressingMode(const DataLayout &DL,
} }
bool SITargetLowering::canMergeStoresTo(unsigned AS, EVT MemVT, bool SITargetLowering::canMergeStoresTo(unsigned AS, EVT MemVT,
const SelectionDAG &DAG) const { const MachineFunction &MF) const {
if (AS == AMDGPUAS::GLOBAL_ADDRESS || AS == AMDGPUAS::FLAT_ADDRESS) { if (AS == AMDGPUAS::GLOBAL_ADDRESS || AS == AMDGPUAS::FLAT_ADDRESS) {
return (MemVT.getSizeInBits() <= 4 * 32); return (MemVT.getSizeInBits() <= 4 * 32);
} else if (AS == AMDGPUAS::PRIVATE_ADDRESS) { } else if (AS == AMDGPUAS::PRIVATE_ADDRESS) {

View File

@ -16,6 +16,7 @@
#include "AMDGPUISelLowering.h" #include "AMDGPUISelLowering.h"
#include "AMDGPUArgumentUsageInfo.h" #include "AMDGPUArgumentUsageInfo.h"
#include "llvm/CodeGen/MachineFunction.h"
namespace llvm { namespace llvm {
@ -267,7 +268,7 @@ public:
Instruction *I = nullptr) const override; Instruction *I = nullptr) const override;
bool canMergeStoresTo(unsigned AS, EVT MemVT, bool canMergeStoresTo(unsigned AS, EVT MemVT,
const SelectionDAG &DAG) const override; const MachineFunction &MF) const override;
bool allowsMisalignedMemoryAccessesImpl( bool allowsMisalignedMemoryAccessesImpl(
unsigned Size, unsigned AddrSpace, Align Alignment, unsigned Size, unsigned AddrSpace, Align Alignment,

View File

@ -680,7 +680,7 @@ class VectorType;
unsigned &Cost) const override; unsigned &Cost) const override;
bool canMergeStoresTo(unsigned AddressSpace, EVT MemVT, bool canMergeStoresTo(unsigned AddressSpace, EVT MemVT,
const SelectionDAG &DAG) const override { const MachineFunction &MF) const override {
// Do not merge to larger than i32. // Do not merge to larger than i32.
return (MemVT.getSizeInBits() <= 32); return (MemVT.getSizeInBits() <= 32);
} }

View File

@ -5387,11 +5387,10 @@ bool X86TargetLowering::isLoadBitCastBeneficial(EVT LoadVT, EVT BitcastVT,
} }
bool X86TargetLowering::canMergeStoresTo(unsigned AddressSpace, EVT MemVT, bool X86TargetLowering::canMergeStoresTo(unsigned AddressSpace, EVT MemVT,
const SelectionDAG &DAG) const { const MachineFunction &MF) const {
// Do not merge to float value size (128 bytes) if no implicit // Do not merge to float value size (128 bytes) if no implicit
// float attribute is set. // float attribute is set.
bool NoFloat = DAG.getMachineFunction().getFunction().hasFnAttribute( bool NoFloat = MF.getFunction().hasFnAttribute(Attribute::NoImplicitFloat);
Attribute::NoImplicitFloat);
if (NoFloat) { if (NoFloat) {
unsigned MaxIntSize = Subtarget.is64Bit() ? 64 : 32; unsigned MaxIntSize = Subtarget.is64Bit() ? 64 : 32;

View File

@ -14,6 +14,7 @@
#ifndef LLVM_LIB_TARGET_X86_X86ISELLOWERING_H #ifndef LLVM_LIB_TARGET_X86_X86ISELLOWERING_H
#define LLVM_LIB_TARGET_X86_X86ISELLOWERING_H #define LLVM_LIB_TARGET_X86_X86ISELLOWERING_H
#include "llvm/CodeGen/MachineFunction.h"
#include "llvm/CodeGen/TargetLowering.h" #include "llvm/CodeGen/TargetLowering.h"
namespace llvm { namespace llvm {
@ -989,7 +990,7 @@ namespace llvm {
} }
bool canMergeStoresTo(unsigned AddressSpace, EVT MemVT, bool canMergeStoresTo(unsigned AddressSpace, EVT MemVT,
const SelectionDAG &DAG) const override; const MachineFunction &MF) const override;
bool isCheapToSpeculateCttz() const override; bool isCheapToSpeculateCttz() const override;