2012-02-18 20:03:15 +08:00
|
|
|
//===-- PPCInstrInfo.h - PowerPC Instruction Information --------*- C++ -*-===//
|
2005-04-22 07:30:14 +08:00
|
|
|
//
|
2019-01-19 16:50:56 +08:00
|
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
2005-04-22 07:30:14 +08:00
|
|
|
//
|
2004-08-17 12:55:41 +08:00
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
|
|
|
// This file contains the PowerPC implementation of the TargetInstrInfo class.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2014-08-14 00:26:38 +08:00
|
|
|
#ifndef LLVM_LIB_TARGET_POWERPC_PPCINSTRINFO_H
|
|
|
|
#define LLVM_LIB_TARGET_POWERPC_PPCINSTRINFO_H
|
2004-08-17 12:55:41 +08:00
|
|
|
|
2005-10-15 07:59:06 +08:00
|
|
|
#include "PPCRegisterInfo.h"
|
2017-11-08 09:01:31 +08:00
|
|
|
#include "llvm/CodeGen/TargetInstrInfo.h"
|
2004-08-17 12:55:41 +08:00
|
|
|
|
2011-07-02 01:57:27 +08:00
|
|
|
#define GET_INSTRINFO_HEADER
|
|
|
|
#include "PPCGenInstrInfo.inc"
|
|
|
|
|
2004-08-17 12:55:41 +08:00
|
|
|
namespace llvm {
|
2006-03-12 17:13:49 +08:00
|
|
|
|
|
|
|
/// PPCII - This namespace holds all of the PowerPC target-specific
|
|
|
|
/// per-instruction flags. These must match the corresponding definitions in
|
|
|
|
/// PPC.td and PPCInstrFormats.td.
|
|
|
|
namespace PPCII {
|
|
|
|
enum {
|
|
|
|
// PPC970 Instruction Flags. These flags describe the characteristics of the
|
|
|
|
// PowerPC 970 (aka G5) dispatch groups and how they are formed out of
|
|
|
|
// raw machine instructions.
|
|
|
|
|
|
|
|
/// PPC970_First - This instruction starts a new dispatch group, so it will
|
|
|
|
/// always be the first one in the group.
|
|
|
|
PPC970_First = 0x1,
|
2010-12-24 12:28:06 +08:00
|
|
|
|
2006-03-12 17:13:49 +08:00
|
|
|
/// PPC970_Single - This instruction starts a new dispatch group and
|
|
|
|
/// terminates it, so it will be the sole instruction in the group.
|
|
|
|
PPC970_Single = 0x2,
|
|
|
|
|
2006-03-13 13:15:10 +08:00
|
|
|
/// PPC970_Cracked - This instruction is cracked into two pieces, requiring
|
|
|
|
/// two dispatch pipes to be available to issue.
|
|
|
|
PPC970_Cracked = 0x4,
|
2010-12-24 12:28:06 +08:00
|
|
|
|
2006-03-12 17:13:49 +08:00
|
|
|
/// PPC970_Mask/Shift - This is a bitmask that selects the pipeline type that
|
|
|
|
/// an instruction is issued to.
|
2006-03-13 13:15:10 +08:00
|
|
|
PPC970_Shift = 3,
|
2006-05-25 01:04:05 +08:00
|
|
|
PPC970_Mask = 0x07 << PPC970_Shift
|
2006-03-12 17:13:49 +08:00
|
|
|
};
|
|
|
|
enum PPC970_Unit {
|
|
|
|
/// These are the various PPC970 execution unit pipelines. Each instruction
|
|
|
|
/// is one of these.
|
|
|
|
PPC970_Pseudo = 0 << PPC970_Shift, // Pseudo instruction
|
|
|
|
PPC970_FXU = 1 << PPC970_Shift, // Fixed Point (aka Integer/ALU) Unit
|
|
|
|
PPC970_LSU = 2 << PPC970_Shift, // Load Store Unit
|
|
|
|
PPC970_FPU = 3 << PPC970_Shift, // Floating Point Unit
|
|
|
|
PPC970_CRU = 4 << PPC970_Shift, // Control Register Unit
|
|
|
|
PPC970_VALU = 5 << PPC970_Shift, // Vector ALU
|
|
|
|
PPC970_VPERM = 6 << PPC970_Shift, // Vector Permute Unit
|
2006-05-25 01:04:05 +08:00
|
|
|
PPC970_BRU = 7 << PPC970_Shift // Branch Unit
|
2006-03-12 17:13:49 +08:00
|
|
|
};
|
2016-10-04 14:59:23 +08:00
|
|
|
|
|
|
|
enum {
|
|
|
|
/// Shift count to bypass PPC970 flags
|
|
|
|
NewDef_Shift = 6,
|
|
|
|
|
2018-03-27 01:39:18 +08:00
|
|
|
/// This instruction is an X-Form memory operation.
|
2020-01-24 04:01:02 +08:00
|
|
|
XFormMemOp = 0x1 << NewDef_Shift,
|
|
|
|
/// This instruction is prefixed.
|
|
|
|
Prefixed = 0x1 << (NewDef_Shift+1)
|
2016-10-04 14:59:23 +08:00
|
|
|
};
|
2010-11-15 07:42:06 +08:00
|
|
|
} // end namespace PPCII
|
2010-12-24 12:28:06 +08:00
|
|
|
|
2017-12-15 15:27:53 +08:00
|
|
|
// Instructions that have an immediate form might be convertible to that
|
|
|
|
// form if the correct input is a result of a load immediate. In order to
|
|
|
|
// know whether the transformation is special, we might need to know some
|
|
|
|
// of the details of the two forms.
|
|
|
|
struct ImmInstrInfo {
|
|
|
|
// Is the immediate field in the immediate form signed or unsigned?
|
|
|
|
uint64_t SignedImm : 1;
|
|
|
|
// Does the immediate need to be a multiple of some value?
|
|
|
|
uint64_t ImmMustBeMultipleOf : 5;
|
|
|
|
// Is R0/X0 treated specially by the original r+r instruction?
|
|
|
|
// If so, in which operand?
|
|
|
|
uint64_t ZeroIsSpecialOrig : 3;
|
|
|
|
// Is R0/X0 treated specially by the new r+i instruction?
|
|
|
|
// If so, in which operand?
|
|
|
|
uint64_t ZeroIsSpecialNew : 3;
|
|
|
|
// Is the operation commutative?
|
|
|
|
uint64_t IsCommutative : 1;
|
2018-08-20 10:52:55 +08:00
|
|
|
// The operand number to check for add-immediate def.
|
|
|
|
uint64_t OpNoForForwarding : 3;
|
2017-12-15 15:27:53 +08:00
|
|
|
// The operand number for the immediate.
|
|
|
|
uint64_t ImmOpNo : 3;
|
|
|
|
// The opcode of the new instruction.
|
|
|
|
uint64_t ImmOpcode : 16;
|
|
|
|
// The size of the immediate.
|
|
|
|
uint64_t ImmWidth : 5;
|
2017-12-29 20:22:27 +08:00
|
|
|
// The immediate should be truncated to N bits.
|
|
|
|
uint64_t TruncateImmTo : 5;
|
2018-08-20 10:52:55 +08:00
|
|
|
// Is the instruction summing the operand
|
|
|
|
uint64_t IsSummingOperands : 1;
|
2017-12-15 15:27:53 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
// Information required to convert an instruction to just a materialized
|
|
|
|
// immediate.
|
|
|
|
struct LoadImmediateInfo {
|
|
|
|
unsigned Imm : 16;
|
|
|
|
unsigned Is64Bit : 1;
|
|
|
|
unsigned SetCR : 1;
|
|
|
|
};
|
|
|
|
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
// Index into the OpcodesForSpill array.
|
|
|
|
enum SpillOpcodeKey {
|
|
|
|
SOK_Int4Spill,
|
|
|
|
SOK_Int8Spill,
|
|
|
|
SOK_Float8Spill,
|
|
|
|
SOK_Float4Spill,
|
|
|
|
SOK_CRSpill,
|
|
|
|
SOK_CRBitSpill,
|
|
|
|
SOK_VRVectorSpill,
|
|
|
|
SOK_VSXVectorSpill,
|
|
|
|
SOK_VectorFloat8Spill,
|
|
|
|
SOK_VectorFloat4Spill,
|
|
|
|
SOK_SpillToVSR,
|
2020-11-02 23:38:24 +08:00
|
|
|
SOK_PairedVecSpill,
|
|
|
|
SOK_AccumulatorSpill,
|
|
|
|
SOK_UAccumulatorSpill,
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
SOK_SPESpill,
|
|
|
|
SOK_LastOpcodeSpill // This must be last on the enum.
|
|
|
|
};
|
|
|
|
|
|
|
|
// Define list of load and store spill opcodes.
|
2020-11-02 23:38:24 +08:00
|
|
|
#define NoInstr PPC::INSTRUCTION_LIST_END
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
#define Pwr8LoadOpcodes \
|
|
|
|
{ \
|
|
|
|
PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR, \
|
|
|
|
PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXVD2X, PPC::LXSDX, PPC::LXSSPX, \
|
2020-11-02 23:38:24 +08:00
|
|
|
PPC::SPILLTOVSR_LD, NoInstr, NoInstr, NoInstr, PPC::EVLDD \
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
#define Pwr9LoadOpcodes \
|
|
|
|
{ \
|
|
|
|
PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR, \
|
|
|
|
PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXV, PPC::DFLOADf64, \
|
2020-11-02 23:38:24 +08:00
|
|
|
PPC::DFLOADf32, PPC::SPILLTOVSR_LD, NoInstr, NoInstr, NoInstr, NoInstr \
|
|
|
|
}
|
|
|
|
|
|
|
|
#define Pwr10LoadOpcodes \
|
|
|
|
{ \
|
|
|
|
PPC::LWZ, PPC::LD, PPC::LFD, PPC::LFS, PPC::RESTORE_CR, \
|
|
|
|
PPC::RESTORE_CRBIT, PPC::LVX, PPC::LXV, PPC::DFLOADf64, \
|
|
|
|
PPC::DFLOADf32, PPC::SPILLTOVSR_LD, PPC::LXVP, PPC::RESTORE_ACC, \
|
|
|
|
PPC::RESTORE_UACC, NoInstr \
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
#define Pwr8StoreOpcodes \
|
|
|
|
{ \
|
|
|
|
PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \
|
2020-09-30 21:56:55 +08:00
|
|
|
PPC::STVX, PPC::STXVD2X, PPC::STXSDX, PPC::STXSSPX, \
|
2020-11-02 23:38:24 +08:00
|
|
|
PPC::SPILLTOVSR_ST, NoInstr, NoInstr, NoInstr, PPC::EVSTDD \
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
#define Pwr9StoreOpcodes \
|
|
|
|
{ \
|
|
|
|
PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \
|
|
|
|
PPC::STVX, PPC::STXV, PPC::DFSTOREf64, PPC::DFSTOREf32, \
|
2020-11-02 23:38:24 +08:00
|
|
|
PPC::SPILLTOVSR_ST, NoInstr, NoInstr, NoInstr, NoInstr \
|
|
|
|
}
|
|
|
|
|
|
|
|
#define Pwr10StoreOpcodes \
|
|
|
|
{ \
|
|
|
|
PPC::STW, PPC::STD, PPC::STFD, PPC::STFS, PPC::SPILL_CR, PPC::SPILL_CRBIT, \
|
|
|
|
PPC::STVX, PPC::STXV, PPC::DFSTOREf64, PPC::DFSTOREf32, \
|
|
|
|
PPC::SPILLTOVSR_ST, PPC::STXVP, PPC::SPILL_ACC, PPC::SPILL_UACC, \
|
|
|
|
NoInstr \
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Initialize arrays for load and store spill opcodes on supported subtargets.
|
|
|
|
#define StoreOpcodesForSpill \
|
2020-11-02 23:38:24 +08:00
|
|
|
{ Pwr8StoreOpcodes, Pwr9StoreOpcodes, Pwr10StoreOpcodes }
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
#define LoadOpcodesForSpill \
|
2020-11-02 23:38:24 +08:00
|
|
|
{ Pwr8LoadOpcodes, Pwr9LoadOpcodes, Pwr10LoadOpcodes }
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
|
2015-03-12 14:07:16 +08:00
|
|
|
class PPCSubtarget;
|
2011-07-02 01:57:27 +08:00
|
|
|
class PPCInstrInfo : public PPCGenInstrInfo {
|
2014-06-13 05:48:52 +08:00
|
|
|
PPCSubtarget &Subtarget;
|
2005-10-16 13:39:50 +08:00
|
|
|
const PPCRegisterInfo RI;
|
2020-11-02 23:38:24 +08:00
|
|
|
const unsigned StoreSpillOpcodesArray[3][SOK_LastOpcodeSpill] =
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
StoreOpcodesForSpill;
|
2020-11-02 23:38:24 +08:00
|
|
|
const unsigned LoadSpillOpcodesArray[3][SOK_LastOpcodeSpill] =
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
LoadOpcodesForSpill;
|
2008-03-11 06:49:16 +08:00
|
|
|
|
2018-03-27 01:39:18 +08:00
|
|
|
void StoreRegToStackSlot(MachineFunction &MF, unsigned SrcReg, bool isKill,
|
|
|
|
int FrameIdx, const TargetRegisterClass *RC,
|
|
|
|
SmallVectorImpl<MachineInstr *> &NewMIs) const;
|
|
|
|
void LoadRegFromStackSlot(MachineFunction &MF, const DebugLoc &DL,
|
2008-07-08 07:14:23 +08:00
|
|
|
unsigned DestReg, int FrameIdx,
|
2008-03-11 06:49:16 +08:00
|
|
|
const TargetRegisterClass *RC,
|
2018-03-27 01:39:18 +08:00
|
|
|
SmallVectorImpl<MachineInstr *> &NewMIs) const;
|
2018-08-20 10:52:55 +08:00
|
|
|
|
2020-06-12 14:58:34 +08:00
|
|
|
// Replace the instruction with single LI if possible. \p DefMI must be LI or
|
|
|
|
// LI8.
|
|
|
|
bool simplifyToLI(MachineInstr &MI, MachineInstr &DefMI,
|
|
|
|
unsigned OpNoForForwarding, MachineInstr **KilledDef) const;
|
2020-06-19 11:31:51 +08:00
|
|
|
// If the inst is imm-form and its register operand is produced by a ADDI, put
|
|
|
|
// the imm into the inst directly and remove the ADDI if possible.
|
|
|
|
bool transformToNewImmFormFedByAdd(MachineInstr &MI, MachineInstr &DefMI,
|
|
|
|
unsigned OpNoForForwarding) const;
|
2020-06-12 14:58:34 +08:00
|
|
|
// If the inst is x-form and has imm-form and one of its operand is produced
|
|
|
|
// by a LI, put the imm into the inst directly and remove the LI if possible.
|
2018-08-20 10:52:55 +08:00
|
|
|
bool transformToImmFormFedByLI(MachineInstr &MI, const ImmInstrInfo &III,
|
2020-06-12 14:58:34 +08:00
|
|
|
unsigned ConstantOpNo,
|
|
|
|
MachineInstr &DefMI) const;
|
|
|
|
// If the inst is x-form and has imm-form and one of its operand is produced
|
|
|
|
// by an add-immediate, try to transform it when possible.
|
2018-08-20 10:52:55 +08:00
|
|
|
bool transformToImmFormFedByAdd(MachineInstr &MI, const ImmInstrInfo &III,
|
2019-03-05 12:56:54 +08:00
|
|
|
unsigned ConstantOpNo, MachineInstr &DefMI,
|
2018-08-20 10:52:55 +08:00
|
|
|
bool KillDefMI) const;
|
2018-09-12 01:28:43 +08:00
|
|
|
// Try to find that, if the instruction 'MI' contains any operand that
|
2018-08-20 10:52:55 +08:00
|
|
|
// could be forwarded from some inst that feeds it. If yes, return the
|
|
|
|
// Def of that operand. And OpNoForForwarding is the operand index in
|
|
|
|
// the 'MI' for that 'Def'. If we see another use of this Def between
|
|
|
|
// the Def and the MI, SeenIntermediateUse becomes 'true'.
|
|
|
|
MachineInstr *getForwardingDefMI(MachineInstr &MI,
|
|
|
|
unsigned &OpNoForForwarding,
|
|
|
|
bool &SeenIntermediateUse) const;
|
|
|
|
|
|
|
|
// Can the user MI have it's source at index \p OpNoForForwarding
|
|
|
|
// forwarded from an add-immediate that feeds it?
|
|
|
|
bool isUseMIElgibleForForwarding(MachineInstr &MI, const ImmInstrInfo &III,
|
|
|
|
unsigned OpNoForForwarding) const;
|
|
|
|
bool isDefMIElgibleForForwarding(MachineInstr &DefMI,
|
|
|
|
const ImmInstrInfo &III,
|
|
|
|
MachineOperand *&ImmMO,
|
|
|
|
MachineOperand *&RegMO) const;
|
|
|
|
bool isImmElgibleForForwarding(const MachineOperand &ImmMO,
|
|
|
|
const MachineInstr &DefMI,
|
|
|
|
const ImmInstrInfo &III,
|
2020-06-19 11:31:51 +08:00
|
|
|
int64_t &Imm,
|
|
|
|
int64_t BaseImm = 0) const;
|
2018-08-20 10:52:55 +08:00
|
|
|
bool isRegElgibleForForwarding(const MachineOperand &RegMO,
|
|
|
|
const MachineInstr &DefMI,
|
2019-03-05 12:56:54 +08:00
|
|
|
const MachineInstr &MI, bool KillDefMI,
|
|
|
|
bool &IsFwdFeederRegKilled) const;
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
unsigned getSpillTarget() const;
|
2018-03-27 01:39:18 +08:00
|
|
|
const unsigned *getStoreOpcodesForSpillArray() const;
|
|
|
|
const unsigned *getLoadOpcodesForSpillArray() const;
|
2020-11-02 23:38:24 +08:00
|
|
|
unsigned getSpillIndex(const TargetRegisterClass *RC) const;
|
[PowerPC] fma chain break to expose more ILP
This patch tries to reassociate two patterns related to FMA to expose
more ILP on PowerPC.
// Pattern 1:
// A = FADD X, Y (Leaf)
// B = FMA A, M21, M22 (Prev)
// C = FMA B, M31, M32 (Root)
// -->
// A = FMA X, M21, M22
// B = FMA Y, M31, M32
// C = FADD A, B
// Pattern 2:
// A = FMA X, M11, M12 (Leaf)
// B = FMA A, M21, M22 (Prev)
// C = FMA B, M31, M32 (Root)
// -->
// A = FMUL M11, M12
// B = FMA X, M21, M22
// D = FMA A, M31, M32
// C = FADD B, D
Reviewed By: jsji
Differential Revision: https://reviews.llvm.org/D80175
2020-06-15 09:34:52 +08:00
|
|
|
int16_t getFMAOpIdxInfo(unsigned Opcode) const;
|
|
|
|
void reassociateFMA(MachineInstr &Root, MachineCombinerPattern Pattern,
|
|
|
|
SmallVectorImpl<MachineInstr *> &InsInstrs,
|
|
|
|
SmallVectorImpl<MachineInstr *> &DelInstrs,
|
|
|
|
DenseMap<unsigned, unsigned> &InstrIdxForVirtReg) const;
|
2021-01-18 12:53:33 +08:00
|
|
|
bool isLoadFromConstantPool(MachineInstr *I) const;
|
|
|
|
Register
|
|
|
|
generateLoadForNewConst(unsigned Idx, MachineInstr *MI, Type *Ty,
|
|
|
|
SmallVectorImpl<MachineInstr *> &InsInstrs) const;
|
|
|
|
const Constant *getConstantFromConstantPool(MachineInstr *I) const;
|
2013-11-19 08:57:56 +08:00
|
|
|
virtual void anchor();
|
2015-09-29 04:33:22 +08:00
|
|
|
|
|
|
|
protected:
|
|
|
|
/// Commutes the operands in the given instruction.
|
|
|
|
/// The commutable operands are specified by their indices OpIdx1 and OpIdx2.
|
|
|
|
///
|
|
|
|
/// Do not call this method for a non-commutable instruction or for
|
|
|
|
/// non-commutable pair of operand indices OpIdx1 and OpIdx2.
|
|
|
|
/// Even though the instruction is commutable, the method may still
|
|
|
|
/// fail to commute the operands, null pointer is returned in such cases.
|
|
|
|
///
|
|
|
|
/// For example, we can commute rlwimi instructions, but only if the
|
|
|
|
/// rotate amt is zero. We also have to munge the immediates a bit.
|
2016-06-30 08:01:54 +08:00
|
|
|
MachineInstr *commuteInstructionImpl(MachineInstr &MI, bool NewMI,
|
2015-09-29 04:33:22 +08:00
|
|
|
unsigned OpIdx1,
|
|
|
|
unsigned OpIdx2) const override;
|
|
|
|
|
2004-08-17 12:55:41 +08:00
|
|
|
public:
|
2014-06-13 05:48:52 +08:00
|
|
|
explicit PPCInstrInfo(PPCSubtarget &STI);
|
2004-08-17 12:55:41 +08:00
|
|
|
|
|
|
|
/// getRegisterInfo - TargetInstrInfo is a superset of MRegister info. As
|
|
|
|
/// such, whenever a client has an instance of instruction info, it should
|
|
|
|
/// always be able to get register info as well (through this method).
|
|
|
|
///
|
2014-04-29 15:57:37 +08:00
|
|
|
const PPCRegisterInfo &getRegisterInfo() const { return RI; }
|
2004-08-17 12:55:41 +08:00
|
|
|
|
2018-03-27 01:39:18 +08:00
|
|
|
bool isXFormMemOp(unsigned Opcode) const {
|
|
|
|
return get(Opcode).TSFlags & PPCII::XFormMemOp;
|
|
|
|
}
|
2020-01-28 09:20:53 +08:00
|
|
|
bool isPrefixed(unsigned Opcode) const {
|
|
|
|
return get(Opcode).TSFlags & PPCII::Prefixed;
|
|
|
|
}
|
|
|
|
|
2018-10-09 18:54:04 +08:00
|
|
|
static bool isSameClassPhysRegCopy(unsigned Opcode) {
|
2020-07-28 11:02:20 +08:00
|
|
|
unsigned CopyOpcodes[] = {PPC::OR, PPC::OR8, PPC::FMR,
|
|
|
|
PPC::VOR, PPC::XXLOR, PPC::XXLORf,
|
|
|
|
PPC::XSCPSGNDP, PPC::MCRF, PPC::CROR,
|
|
|
|
PPC::EVOR, -1U};
|
2018-10-09 18:54:04 +08:00
|
|
|
for (int i = 0; CopyOpcodes[i] != -1U; i++)
|
|
|
|
if (Opcode == CopyOpcodes[i])
|
|
|
|
return true;
|
|
|
|
return false;
|
|
|
|
}
|
2018-03-27 01:39:18 +08:00
|
|
|
|
Various bits of framework needed for precise machine-level selection
DAG scheduling during isel. Most new functionality is currently
guarded by -enable-sched-cycles and -enable-sched-hazard.
Added InstrItineraryData::IssueWidth field, currently derived from
ARM itineraries, but could be initialized differently on other targets.
Added ScheduleHazardRecognizer::MaxLookAhead to indicate whether it is
active, and if so how many cycles of state it holds.
Added SchedulingPriorityQueue::HasReadyFilter to allowing gating entry
into the scheduler's available queue.
ScoreboardHazardRecognizer now accesses the ScheduleDAG in order to
get information about it's SUnits, provides RecedeCycle for bottom-up
scheduling, correctly computes scoreboard depth, tracks IssueCount, and
considers potential stall cycles when checking for hazards.
ScheduleDAGRRList now models machine cycles and hazards (under
flags). It tracks MinAvailableCycle, drives the hazard recognizer and
priority queue's ready filter, manages a new PendingQueue, properly
accounts for stall cycles, etc.
llvm-svn: 122541
2010-12-24 13:03:26 +08:00
|
|
|
ScheduleHazardRecognizer *
|
2014-06-14 06:38:52 +08:00
|
|
|
CreateTargetHazardRecognizer(const TargetSubtargetInfo *STI,
|
2014-04-29 15:57:37 +08:00
|
|
|
const ScheduleDAG *DAG) const override;
|
2011-12-02 12:58:02 +08:00
|
|
|
ScheduleHazardRecognizer *
|
|
|
|
CreateTargetPostRAHazardRecognizer(const InstrItineraryData *II,
|
2014-04-29 15:57:37 +08:00
|
|
|
const ScheduleDAG *DAG) const override;
|
Various bits of framework needed for precise machine-level selection
DAG scheduling during isel. Most new functionality is currently
guarded by -enable-sched-cycles and -enable-sched-hazard.
Added InstrItineraryData::IssueWidth field, currently derived from
ARM itineraries, but could be initialized differently on other targets.
Added ScheduleHazardRecognizer::MaxLookAhead to indicate whether it is
active, and if so how many cycles of state it holds.
Added SchedulingPriorityQueue::HasReadyFilter to allowing gating entry
into the scheduler's available queue.
ScoreboardHazardRecognizer now accesses the ScheduleDAG in order to
get information about it's SUnits, provides RecedeCycle for bottom-up
scheduling, correctly computes scoreboard depth, tracks IssueCount, and
considers potential stall cycles when checking for hazards.
ScheduleDAGRRList now models machine cycles and hazards (under
flags). It tracks MinAvailableCycle, drives the hazard recognizer and
priority queue's ready filter, manages a new PendingQueue, properly
accounts for stall cycles, etc.
llvm-svn: 122541
2010-12-24 13:03:26 +08:00
|
|
|
|
[PowerPC] Fix the PPCInstrInfo::getInstrLatency implementation
PowerPC uses itineraries to describe processor pipelines (and dispatch-group
restrictions for P7/P8 cores). Unfortunately, the target-independent
implementation of TII.getInstrLatency calls ItinData->getStageLatency, and that
looks for the largest cycle count in the pipeline for any given instruction.
This, however, yields the wrong answer for the PPC itineraries, because we
don't encode the full pipeline. Because the functional units are fully
pipelined, we only model the initial stages (there are no relevant hazards in
the later stages to model), and so the technique employed by getStageLatency
does not really work. Instead, we should take the maximum output operand
latency, and that's what PPCInstrInfo::getInstrLatency now does.
This caused some test-case churn, including two unfortunate side effects.
First, the new arrangement of copies we get from function parameters now
sometimes blocks VSX FMA mutation (a FIXME has been added to the code and the
test cases), and we have one significant test-suite regression:
SingleSource/Benchmarks/BenchmarkGame/spectral-norm
56.4185% +/- 18.9398%
In this benchmark we have a loop with a vectorized FP divide, and it with the
new scheduling both divides end up in the same dispatch group (which in this
case seems to cause a problem, although why is not exactly clear). The grouping
structure is hard to predict from the bottom of the loop, and there may not be
much we can do to fix this.
Very few other test-suite performance effects were really significant, but
almost all weakly favor this change. However, in light of the issues
highlighted above, I've left the old behavior available via a
command-line flag.
llvm-svn: 242188
2015-07-15 04:02:02 +08:00
|
|
|
unsigned getInstrLatency(const InstrItineraryData *ItinData,
|
2016-06-30 08:01:54 +08:00
|
|
|
const MachineInstr &MI,
|
[PowerPC] Fix the PPCInstrInfo::getInstrLatency implementation
PowerPC uses itineraries to describe processor pipelines (and dispatch-group
restrictions for P7/P8 cores). Unfortunately, the target-independent
implementation of TII.getInstrLatency calls ItinData->getStageLatency, and that
looks for the largest cycle count in the pipeline for any given instruction.
This, however, yields the wrong answer for the PPC itineraries, because we
don't encode the full pipeline. Because the functional units are fully
pipelined, we only model the initial stages (there are no relevant hazards in
the later stages to model), and so the technique employed by getStageLatency
does not really work. Instead, we should take the maximum output operand
latency, and that's what PPCInstrInfo::getInstrLatency now does.
This caused some test-case churn, including two unfortunate side effects.
First, the new arrangement of copies we get from function parameters now
sometimes blocks VSX FMA mutation (a FIXME has been added to the code and the
test cases), and we have one significant test-suite regression:
SingleSource/Benchmarks/BenchmarkGame/spectral-norm
56.4185% +/- 18.9398%
In this benchmark we have a loop with a vectorized FP divide, and it with the
new scheduling both divides end up in the same dispatch group (which in this
case seems to cause a problem, although why is not exactly clear). The grouping
structure is hard to predict from the bottom of the loop, and there may not be
much we can do to fix this.
Very few other test-suite performance effects were really significant, but
almost all weakly favor this change. However, in light of the issues
highlighted above, I've left the old behavior available via a
command-line flag.
llvm-svn: 242188
2015-07-15 04:02:02 +08:00
|
|
|
unsigned *PredCost = nullptr) const override;
|
|
|
|
|
2013-12-12 08:19:11 +08:00
|
|
|
int getOperandLatency(const InstrItineraryData *ItinData,
|
2016-06-30 08:01:54 +08:00
|
|
|
const MachineInstr &DefMI, unsigned DefIdx,
|
|
|
|
const MachineInstr &UseMI,
|
2014-04-29 15:57:37 +08:00
|
|
|
unsigned UseIdx) const override;
|
2013-12-12 08:19:11 +08:00
|
|
|
int getOperandLatency(const InstrItineraryData *ItinData,
|
|
|
|
SDNode *DefNode, unsigned DefIdx,
|
2014-04-29 15:57:37 +08:00
|
|
|
SDNode *UseNode, unsigned UseIdx) const override {
|
2013-12-12 08:19:11 +08:00
|
|
|
return PPCGenInstrInfo::getOperandLatency(ItinData, DefNode, DefIdx,
|
|
|
|
UseNode, UseIdx);
|
|
|
|
}
|
|
|
|
|
2015-06-13 11:42:11 +08:00
|
|
|
bool hasLowDefLatency(const TargetSchedModel &SchedModel,
|
2016-06-30 08:01:54 +08:00
|
|
|
const MachineInstr &DefMI,
|
2015-01-09 06:11:49 +08:00
|
|
|
unsigned DefIdx) const override {
|
|
|
|
// Machine LICM should hoist all instructions in low-register-pressure
|
|
|
|
// situations; none are sufficiently free to justify leaving in a loop
|
|
|
|
// body.
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2015-07-15 16:23:05 +08:00
|
|
|
bool useMachineCombiner() const override {
|
|
|
|
return true;
|
|
|
|
}
|
2015-09-21 23:09:11 +08:00
|
|
|
|
[PowerPC] fma chain break to expose more ILP
This patch tries to reassociate two patterns related to FMA to expose
more ILP on PowerPC.
// Pattern 1:
// A = FADD X, Y (Leaf)
// B = FMA A, M21, M22 (Prev)
// C = FMA B, M31, M32 (Root)
// -->
// A = FMA X, M21, M22
// B = FMA Y, M31, M32
// C = FADD A, B
// Pattern 2:
// A = FMA X, M11, M12 (Leaf)
// B = FMA A, M21, M22 (Prev)
// C = FMA B, M31, M32 (Root)
// -->
// A = FMUL M11, M12
// B = FMA X, M21, M22
// D = FMA A, M31, M32
// C = FADD B, D
Reviewed By: jsji
Differential Revision: https://reviews.llvm.org/D80175
2020-06-15 09:34:52 +08:00
|
|
|
/// When getMachineCombinerPatterns() finds patterns, this function generates
|
|
|
|
/// the instructions that could replace the original code sequence
|
|
|
|
void genAlternativeCodeSequence(
|
|
|
|
MachineInstr &Root, MachineCombinerPattern Pattern,
|
|
|
|
SmallVectorImpl<MachineInstr *> &InsInstrs,
|
|
|
|
SmallVectorImpl<MachineInstr *> &DelInstrs,
|
|
|
|
DenseMap<unsigned, unsigned> &InstrIdxForVirtReg) const override;
|
|
|
|
|
|
|
|
/// Return true when there is potentially a faster code sequence for a fma
|
|
|
|
/// chain ending in \p Root. All potential patterns are output in the \p
|
|
|
|
/// P array.
|
|
|
|
bool getFMAPatterns(MachineInstr &Root,
|
2021-01-18 12:53:33 +08:00
|
|
|
SmallVectorImpl<MachineCombinerPattern> &P,
|
|
|
|
bool DoRegPressureReduce) const;
|
[PowerPC] fma chain break to expose more ILP
This patch tries to reassociate two patterns related to FMA to expose
more ILP on PowerPC.
// Pattern 1:
// A = FADD X, Y (Leaf)
// B = FMA A, M21, M22 (Prev)
// C = FMA B, M31, M32 (Root)
// -->
// A = FMA X, M21, M22
// B = FMA Y, M31, M32
// C = FADD A, B
// Pattern 2:
// A = FMA X, M11, M12 (Leaf)
// B = FMA A, M21, M22 (Prev)
// C = FMA B, M31, M32 (Root)
// -->
// A = FMUL M11, M12
// B = FMA X, M21, M22
// D = FMA A, M31, M32
// C = FADD B, D
Reviewed By: jsji
Differential Revision: https://reviews.llvm.org/D80175
2020-06-15 09:34:52 +08:00
|
|
|
|
2015-07-15 16:23:05 +08:00
|
|
|
/// Return true when there is potentially a faster code sequence
|
|
|
|
/// for an instruction chain ending in <Root>. All potential patterns are
|
|
|
|
/// output in the <Pattern> array.
|
2020-11-25 10:33:13 +08:00
|
|
|
bool getMachineCombinerPatterns(MachineInstr &Root,
|
|
|
|
SmallVectorImpl<MachineCombinerPattern> &P,
|
|
|
|
bool DoRegPressureReduce) const override;
|
2015-09-21 23:09:11 +08:00
|
|
|
|
2021-01-18 12:53:33 +08:00
|
|
|
/// On PowerPC, we leverage machine combiner pass to reduce register pressure
|
|
|
|
/// when the register pressure is high for one BB.
|
|
|
|
/// Return true if register pressure for \p MBB is high and ABI is supported
|
|
|
|
/// to reduce register pressure. Otherwise return false.
|
|
|
|
bool
|
|
|
|
shouldReduceRegisterPressure(MachineBasicBlock *MBB,
|
|
|
|
RegisterClassInfo *RegClassInfo) const override;
|
|
|
|
|
|
|
|
/// Fixup the placeholders we put in genAlternativeCodeSequence() for
|
|
|
|
/// MachineCombiner.
|
|
|
|
void
|
|
|
|
finalizeInsInstrs(MachineInstr &Root, MachineCombinerPattern &P,
|
|
|
|
SmallVectorImpl<MachineInstr *> &InsInstrs) const override;
|
|
|
|
|
2015-09-21 23:09:11 +08:00
|
|
|
bool isAssociativeAndCommutative(const MachineInstr &Inst) const override;
|
2015-07-15 16:23:05 +08:00
|
|
|
|
[PowerPC] fma chain break to expose more ILP
This patch tries to reassociate two patterns related to FMA to expose
more ILP on PowerPC.
// Pattern 1:
// A = FADD X, Y (Leaf)
// B = FMA A, M21, M22 (Prev)
// C = FMA B, M31, M32 (Root)
// -->
// A = FMA X, M21, M22
// B = FMA Y, M31, M32
// C = FADD A, B
// Pattern 2:
// A = FMA X, M11, M12 (Leaf)
// B = FMA A, M21, M22 (Prev)
// C = FMA B, M31, M32 (Root)
// -->
// A = FMUL M11, M12
// B = FMA X, M21, M22
// D = FMA A, M31, M32
// C = FADD B, D
Reviewed By: jsji
Differential Revision: https://reviews.llvm.org/D80175
2020-06-15 09:34:52 +08:00
|
|
|
/// On PowerPC, we try to reassociate FMA chain which will increase
|
|
|
|
/// instruction size. Set extension resource length limit to 1 for edge case.
|
|
|
|
/// Resource Length is calculated by scaled resource usage in getCycles().
|
|
|
|
/// Because of the division in getCycles(), it returns different cycles due to
|
|
|
|
/// legacy scaled resource usage. So new resource length may be same with
|
|
|
|
/// legacy or 1 bigger than legacy.
|
|
|
|
/// We need to execlude the 1 bigger case even the resource length is not
|
|
|
|
/// perserved for more FMA chain reassociations on PowerPC.
|
|
|
|
int getExtendResourceLenLimit() const override { return 1; }
|
|
|
|
|
2020-04-28 15:31:12 +08:00
|
|
|
void setSpecialOperandAttr(MachineInstr &OldMI1, MachineInstr &OldMI2,
|
|
|
|
MachineInstr &NewMI1,
|
|
|
|
MachineInstr &NewMI2) const override;
|
|
|
|
|
2020-05-19 09:44:47 +08:00
|
|
|
void setSpecialOperandAttr(MachineInstr &MI, uint16_t Flags) const override;
|
|
|
|
|
2012-06-20 05:14:34 +08:00
|
|
|
bool isCoalescableExtInstr(const MachineInstr &MI,
|
2020-04-04 01:22:51 +08:00
|
|
|
Register &SrcReg, Register &DstReg,
|
2014-04-29 15:57:37 +08:00
|
|
|
unsigned &SubIdx) const override;
|
2016-06-30 08:01:54 +08:00
|
|
|
unsigned isLoadFromStackSlot(const MachineInstr &MI,
|
2014-04-29 15:57:37 +08:00
|
|
|
int &FrameIndex) const override;
|
2017-06-22 01:17:56 +08:00
|
|
|
bool isReallyTriviallyReMaterializable(const MachineInstr &MI,
|
2019-10-19 08:22:07 +08:00
|
|
|
AAResults *AA) const override;
|
2016-06-30 08:01:54 +08:00
|
|
|
unsigned isStoreToStackSlot(const MachineInstr &MI,
|
2014-04-29 15:57:37 +08:00
|
|
|
int &FrameIndex) const override;
|
2006-02-03 04:12:32 +08:00
|
|
|
|
2019-09-25 22:55:57 +08:00
|
|
|
bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1,
|
2014-04-29 15:57:37 +08:00
|
|
|
unsigned &SrcOpIdx2) const override;
|
2014-03-26 03:26:43 +08:00
|
|
|
|
2014-04-29 15:57:37 +08:00
|
|
|
void insertNoop(MachineBasicBlock &MBB,
|
|
|
|
MachineBasicBlock::iterator MI) const override;
|
2006-03-06 07:49:55 +08:00
|
|
|
|
2006-10-14 05:21:17 +08:00
|
|
|
|
|
|
|
// Branch analysis.
|
2016-07-15 22:41:04 +08:00
|
|
|
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB,
|
2014-04-29 15:57:37 +08:00
|
|
|
MachineBasicBlock *&FBB,
|
|
|
|
SmallVectorImpl<MachineOperand> &Cond,
|
|
|
|
bool AllowModify) const override;
|
2016-09-15 04:43:16 +08:00
|
|
|
unsigned removeBranch(MachineBasicBlock &MBB,
|
2016-09-15 01:23:48 +08:00
|
|
|
int *BytesRemoved = nullptr) const override;
|
2016-09-15 01:24:15 +08:00
|
|
|
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB,
|
2015-06-12 03:30:37 +08:00
|
|
|
MachineBasicBlock *FBB, ArrayRef<MachineOperand> Cond,
|
2016-09-15 01:23:48 +08:00
|
|
|
const DebugLoc &DL,
|
|
|
|
int *BytesAdded = nullptr) const override;
|
2013-04-06 07:29:01 +08:00
|
|
|
|
|
|
|
// Select analysis.
|
2015-06-12 03:30:37 +08:00
|
|
|
bool canInsertSelect(const MachineBasicBlock &, ArrayRef<MachineOperand> Cond,
|
2020-04-04 01:22:51 +08:00
|
|
|
Register, Register, Register, int &, int &,
|
2020-01-18 06:34:26 +08:00
|
|
|
int &) const override;
|
2015-06-12 03:30:37 +08:00
|
|
|
void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI,
|
2020-04-04 01:22:51 +08:00
|
|
|
const DebugLoc &DL, Register DstReg,
|
|
|
|
ArrayRef<MachineOperand> Cond, Register TrueReg,
|
|
|
|
Register FalseReg) const override;
|
2014-04-29 15:57:37 +08:00
|
|
|
|
2016-06-12 23:39:02 +08:00
|
|
|
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I,
|
2019-11-11 16:24:21 +08:00
|
|
|
const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg,
|
2014-04-29 15:57:37 +08:00
|
|
|
bool KillSrc) const override;
|
|
|
|
|
|
|
|
void storeRegToStackSlot(MachineBasicBlock &MBB,
|
|
|
|
MachineBasicBlock::iterator MBBI,
|
[NFC] unsigned->Register in storeRegTo/loadRegFromStack
Summary:
This patch makes progress on the 'unsigned -> Register' rewrite for
`TargetInstrInfo::loadRegFromStack` and `TII::storeRegToStack`.
Reviewers: arsenm, craig.topper, uweigand, jpienaar, atanasyan, venkatra, robertlytton, dylanmckay, t.p.northover, kparzysz, tstellar, k-ishizaka
Reviewed By: arsenm
Subscribers: wuzish, merge_guards_bot, jyknight, sdardis, nemanjai, jvesely, wdng, nhaehnle, hiraditya, kbarton, fedor.sergeev, asb, rbar, johnrusso, simoncook, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, rogfer01, MartinMosbeck, brucehoult, the_o, PkmX, jocewei, jsji, Jim, lenary, s.egerton, pzheng, sameer.abuasal, apazos, luismarques, kerbowa, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73870
2020-02-03 21:22:06 +08:00
|
|
|
Register SrcReg, bool isKill, int FrameIndex,
|
2014-04-29 15:57:37 +08:00
|
|
|
const TargetRegisterClass *RC,
|
|
|
|
const TargetRegisterInfo *TRI) const override;
|
|
|
|
|
2020-02-08 04:26:11 +08:00
|
|
|
// Emits a register spill without updating the register class for vector
|
|
|
|
// registers. This ensures that when we spill a vector register the
|
|
|
|
// element order in the register is the same as it was in memory.
|
|
|
|
void storeRegToStackSlotNoUpd(MachineBasicBlock &MBB,
|
|
|
|
MachineBasicBlock::iterator MBBI,
|
|
|
|
unsigned SrcReg, bool isKill, int FrameIndex,
|
|
|
|
const TargetRegisterClass *RC,
|
|
|
|
const TargetRegisterInfo *TRI) const;
|
|
|
|
|
2014-04-29 15:57:37 +08:00
|
|
|
void loadRegFromStackSlot(MachineBasicBlock &MBB,
|
|
|
|
MachineBasicBlock::iterator MBBI,
|
[NFC] unsigned->Register in storeRegTo/loadRegFromStack
Summary:
This patch makes progress on the 'unsigned -> Register' rewrite for
`TargetInstrInfo::loadRegFromStack` and `TII::storeRegToStack`.
Reviewers: arsenm, craig.topper, uweigand, jpienaar, atanasyan, venkatra, robertlytton, dylanmckay, t.p.northover, kparzysz, tstellar, k-ishizaka
Reviewed By: arsenm
Subscribers: wuzish, merge_guards_bot, jyknight, sdardis, nemanjai, jvesely, wdng, nhaehnle, hiraditya, kbarton, fedor.sergeev, asb, rbar, johnrusso, simoncook, sabuasal, niosHD, jrtc27, MaskRay, zzheng, edward-jones, rogfer01, MartinMosbeck, brucehoult, the_o, PkmX, jocewei, jsji, Jim, lenary, s.egerton, pzheng, sameer.abuasal, apazos, luismarques, kerbowa, llvm-commits
Tags: #llvm
Differential Revision: https://reviews.llvm.org/D73870
2020-02-03 21:22:06 +08:00
|
|
|
Register DestReg, int FrameIndex,
|
2014-04-29 15:57:37 +08:00
|
|
|
const TargetRegisterClass *RC,
|
|
|
|
const TargetRegisterInfo *TRI) const override;
|
|
|
|
|
2020-02-08 04:26:11 +08:00
|
|
|
// Emits a register reload without updating the register class for vector
|
|
|
|
// registers. This ensures that when we reload a vector register the
|
|
|
|
// element order in the register is the same as it was in memory.
|
|
|
|
void loadRegFromStackSlotNoUpd(MachineBasicBlock &MBB,
|
|
|
|
MachineBasicBlock::iterator MBBI,
|
|
|
|
unsigned DestReg, int FrameIndex,
|
|
|
|
const TargetRegisterClass *RC,
|
|
|
|
const TargetRegisterInfo *TRI) const;
|
|
|
|
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
unsigned getStoreOpcodeForSpill(const TargetRegisterClass *RC) const;
|
2018-03-27 01:39:18 +08:00
|
|
|
|
[PowerPC][NFC] Cleanup load/store spilling code
Summary: Cleanup and commonize code used for spilling to the stack.
Reviewers: stefanp, nemanjai, #powerpc, kamaub
Reviewed By: nemanjai, #powerpc, kamaub
Subscribers: kamaub, hiraditya, wuzish, shchenz, llvm-commits, kbarton
Tags: #llvm, #powerpc
Differential Revision: https://reviews.llvm.org/D79736
2020-05-12 02:49:50 +08:00
|
|
|
unsigned getLoadOpcodeForSpill(const TargetRegisterClass *RC) const;
|
2018-03-27 01:39:18 +08:00
|
|
|
|
2014-04-29 15:57:37 +08:00
|
|
|
bool
|
2016-09-15 04:43:16 +08:00
|
|
|
reverseBranchCondition(SmallVectorImpl<MachineOperand> &Cond) const override;
|
2014-04-29 15:57:37 +08:00
|
|
|
|
2020-04-04 01:22:51 +08:00
|
|
|
bool FoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI, Register Reg,
|
2016-06-30 08:01:54 +08:00
|
|
|
MachineRegisterInfo *MRI) const override;
|
2013-04-07 03:30:30 +08:00
|
|
|
|
2020-05-13 02:15:06 +08:00
|
|
|
bool onlyFoldImmediate(MachineInstr &UseMI, MachineInstr &DefMI,
|
|
|
|
Register Reg) const;
|
|
|
|
|
2013-04-10 06:58:37 +08:00
|
|
|
// If conversion by predication (only supported by some branch instructions).
|
|
|
|
// All of the profitability checks always return true; it is always
|
|
|
|
// profitable to use the predicated branches.
|
2014-04-29 15:57:37 +08:00
|
|
|
bool isProfitableToIfCvt(MachineBasicBlock &MBB,
|
|
|
|
unsigned NumCycles, unsigned ExtraPredCycles,
|
2015-09-11 07:10:42 +08:00
|
|
|
BranchProbability Probability) const override {
|
2013-04-10 06:58:37 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-04-29 15:57:37 +08:00
|
|
|
bool isProfitableToIfCvt(MachineBasicBlock &TMBB,
|
|
|
|
unsigned NumT, unsigned ExtraT,
|
|
|
|
MachineBasicBlock &FMBB,
|
|
|
|
unsigned NumF, unsigned ExtraF,
|
2015-09-11 07:10:42 +08:00
|
|
|
BranchProbability Probability) const override;
|
2013-04-10 06:58:37 +08:00
|
|
|
|
2015-09-11 07:10:42 +08:00
|
|
|
bool isProfitableToDupForIfCvt(MachineBasicBlock &MBB, unsigned NumCycles,
|
|
|
|
BranchProbability Probability) const override {
|
2013-04-10 06:58:37 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2014-04-29 15:57:37 +08:00
|
|
|
bool isProfitableToUnpredicate(MachineBasicBlock &TMBB,
|
|
|
|
MachineBasicBlock &FMBB) const override {
|
2013-04-10 06:58:37 +08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Predication support.
|
2016-02-23 10:46:52 +08:00
|
|
|
bool isPredicated(const MachineInstr &MI) const override;
|
2013-04-10 06:58:37 +08:00
|
|
|
|
2020-08-10 18:25:01 +08:00
|
|
|
bool isSchedulingBoundary(const MachineInstr &MI,
|
|
|
|
const MachineBasicBlock *MBB,
|
|
|
|
const MachineFunction &MF) const override;
|
|
|
|
|
2016-02-23 10:46:52 +08:00
|
|
|
bool PredicateInstruction(MachineInstr &MI,
|
2015-06-12 03:30:37 +08:00
|
|
|
ArrayRef<MachineOperand> Pred) const override;
|
2013-04-10 06:58:37 +08:00
|
|
|
|
2015-06-12 03:30:37 +08:00
|
|
|
bool SubsumesPredicate(ArrayRef<MachineOperand> Pred1,
|
|
|
|
ArrayRef<MachineOperand> Pred2) const override;
|
2013-04-10 06:58:37 +08:00
|
|
|
|
2020-09-28 23:49:41 +08:00
|
|
|
bool ClobbersPredicate(MachineInstr &MI, std::vector<MachineOperand> &Pred,
|
|
|
|
bool SkipDead) const override;
|
2013-04-10 06:58:37 +08:00
|
|
|
|
2013-04-19 06:15:08 +08:00
|
|
|
// Comparison optimization.
|
|
|
|
|
2020-04-04 01:22:51 +08:00
|
|
|
bool analyzeCompare(const MachineInstr &MI, Register &SrcReg,
|
|
|
|
Register &SrcReg2, int &Mask, int &Value) const override;
|
2013-04-19 06:15:08 +08:00
|
|
|
|
2020-04-04 01:22:51 +08:00
|
|
|
bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg,
|
|
|
|
Register SrcReg2, int Mask, int Value,
|
2014-04-29 15:57:37 +08:00
|
|
|
const MachineRegisterInfo *MRI) const override;
|
2019-07-02 11:28:52 +08:00
|
|
|
|
|
|
|
|
|
|
|
/// Return true if get the base operand, byte offset of an instruction and
|
|
|
|
/// the memory width. Width is the size of memory that is being
|
|
|
|
/// loaded/stored (e.g. 1, 2, 4, 8).
|
|
|
|
bool getMemOperandWithOffsetWidth(const MachineInstr &LdSt,
|
|
|
|
const MachineOperand *&BaseOp,
|
|
|
|
int64_t &Offset, unsigned &Width,
|
|
|
|
const TargetRegisterInfo *TRI) const;
|
|
|
|
|
2020-09-13 19:39:49 +08:00
|
|
|
/// Get the base operand and byte offset of an instruction that reads/writes
|
|
|
|
/// memory.
|
|
|
|
bool getMemOperandsWithOffsetWidth(
|
2020-09-14 04:08:17 +08:00
|
|
|
const MachineInstr &LdSt,
|
|
|
|
SmallVectorImpl<const MachineOperand *> &BaseOps, int64_t &Offset,
|
|
|
|
bool &OffsetIsScalable, unsigned &Width,
|
2020-09-13 19:39:49 +08:00
|
|
|
const TargetRegisterInfo *TRI) const override;
|
|
|
|
|
|
|
|
/// Returns true if the two given memory operations should be scheduled
|
|
|
|
/// adjacent.
|
|
|
|
bool shouldClusterMemOps(ArrayRef<const MachineOperand *> BaseOps1,
|
|
|
|
ArrayRef<const MachineOperand *> BaseOps2,
|
|
|
|
unsigned NumLoads, unsigned NumBytes) const override;
|
|
|
|
|
2019-07-02 11:28:52 +08:00
|
|
|
/// Return true if two MIs access different memory addresses and false
|
|
|
|
/// otherwise
|
|
|
|
bool
|
|
|
|
areMemAccessesTriviallyDisjoint(const MachineInstr &MIa,
|
2019-09-27 06:53:44 +08:00
|
|
|
const MachineInstr &MIb) const override;
|
2013-04-19 06:15:08 +08:00
|
|
|
|
2008-04-17 04:10:13 +08:00
|
|
|
/// GetInstSize - Return the number of bytes of code the specified
|
|
|
|
/// instruction may be. This returns the maximum number of bytes.
|
|
|
|
///
|
2016-07-29 16:16:16 +08:00
|
|
|
unsigned getInstSizeInBytes(const MachineInstr &MI) const override;
|
2014-08-09 03:13:23 +08:00
|
|
|
|
2017-04-22 05:48:41 +08:00
|
|
|
void getNoop(MCInst &NopInst) const override;
|
2015-08-30 15:50:35 +08:00
|
|
|
|
|
|
|
std::pair<unsigned, unsigned>
|
|
|
|
decomposeMachineOperandsTargetFlags(unsigned TF) const override;
|
|
|
|
|
|
|
|
ArrayRef<std::pair<unsigned, const char *>>
|
|
|
|
getSerializableDirectMachineOperandTargetFlags() const override;
|
|
|
|
|
|
|
|
ArrayRef<std::pair<unsigned, const char *>>
|
|
|
|
getSerializableBitmaskMachineOperandTargetFlags() const override;
|
2016-04-20 04:14:52 +08:00
|
|
|
|
2017-11-20 22:38:30 +08:00
|
|
|
// Expand VSX Memory Pseudo instruction to either a VSX or a FP instruction.
|
|
|
|
bool expandVSXMemPseudo(MachineInstr &MI) const;
|
|
|
|
|
2016-04-20 04:14:52 +08:00
|
|
|
// Lower pseudo instructions after register allocation.
|
2016-06-30 08:01:54 +08:00
|
|
|
bool expandPostRAPseudo(MachineInstr &MI) const override;
|
2016-10-04 14:59:23 +08:00
|
|
|
|
|
|
|
static bool isVFRegister(unsigned Reg) {
|
|
|
|
return Reg >= PPC::VF0 && Reg <= PPC::VF31;
|
|
|
|
}
|
|
|
|
static bool isVRRegister(unsigned Reg) {
|
|
|
|
return Reg >= PPC::V0 && Reg <= PPC::V31;
|
|
|
|
}
|
|
|
|
const TargetRegisterClass *updatedRC(const TargetRegisterClass *RC) const;
|
2017-05-31 13:40:25 +08:00
|
|
|
static int getRecordFormOpcode(unsigned Opcode);
|
2017-10-16 12:12:57 +08:00
|
|
|
|
2017-11-28 04:26:36 +08:00
|
|
|
bool isTOCSaveMI(const MachineInstr &MI) const;
|
|
|
|
|
2017-10-16 12:12:57 +08:00
|
|
|
bool isSignOrZeroExtended(const MachineInstr &MI, bool SignExt,
|
|
|
|
const unsigned PhiDepth) const;
|
|
|
|
|
|
|
|
/// Return true if the output of the instruction is always a sign-extended,
|
|
|
|
/// i.e. 0 to 31-th bits are same as 32-th bit.
|
|
|
|
bool isSignExtended(const MachineInstr &MI, const unsigned depth = 0) const {
|
|
|
|
return isSignOrZeroExtended(MI, true, depth);
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Return true if the output of the instruction is always zero-extended,
|
|
|
|
/// i.e. 0 to 31-th bits are all zeros
|
|
|
|
bool isZeroExtended(const MachineInstr &MI, const unsigned depth = 0) const {
|
|
|
|
return isSignOrZeroExtended(MI, false, depth);
|
|
|
|
}
|
2017-12-15 15:27:53 +08:00
|
|
|
|
|
|
|
bool convertToImmediateForm(MachineInstr &MI,
|
|
|
|
MachineInstr **KilledDef = nullptr) const;
|
2019-10-25 16:13:30 +08:00
|
|
|
bool foldFrameOffset(MachineInstr &MI) const;
|
2020-12-17 01:12:24 +08:00
|
|
|
bool combineRLWINM(MachineInstr &MI, MachineInstr **ToErase = nullptr) const;
|
2019-10-25 16:13:30 +08:00
|
|
|
bool isADDIInstrEligibleForFolding(MachineInstr &ADDIMI, int64_t &Imm) const;
|
|
|
|
bool isADDInstrEligibleForFolding(MachineInstr &ADDMI) const;
|
|
|
|
bool isImmInstrEligibleForFolding(MachineInstr &MI, unsigned &BaseReg,
|
|
|
|
unsigned &XFormOpcode,
|
|
|
|
int64_t &OffsetOfImmInstr,
|
|
|
|
ImmInstrInfo &III) const;
|
|
|
|
bool isValidToBeChangedReg(MachineInstr *ADDMI, unsigned Index,
|
|
|
|
MachineInstr *&ADDIMI, int64_t &OffsetAddi,
|
|
|
|
int64_t OffsetImm) const;
|
2019-03-05 12:56:54 +08:00
|
|
|
|
|
|
|
/// Fixup killed/dead flag for register \p RegNo between instructions [\p
|
2020-07-22 18:01:52 +08:00
|
|
|
/// StartMI, \p EndMI]. Some pre-RA or post-RA transformations may violate
|
|
|
|
/// register killed/dead flags semantics, this function can be called to fix
|
|
|
|
/// up. Before calling this function,
|
2019-03-05 12:56:54 +08:00
|
|
|
/// 1. Ensure that \p RegNo liveness is killed after instruction \p EndMI.
|
|
|
|
/// 2. Ensure that there is no new definition between (\p StartMI, \p EndMI)
|
2020-08-10 22:55:22 +08:00
|
|
|
/// and possible definition for \p RegNo is \p StartMI or \p EndMI. For
|
|
|
|
/// pre-RA cases, definition may be \p StartMI through COPY, \p StartMI
|
|
|
|
/// will be adjust to true definition.
|
2020-07-22 18:01:52 +08:00
|
|
|
/// 3. We can do accurate fixup for the case when all instructions between
|
|
|
|
/// [\p StartMI, \p EndMI] are in same basic block.
|
|
|
|
/// 4. For the case when \p StartMI and \p EndMI are not in same basic block,
|
|
|
|
/// we conservatively clear kill flag for all uses of \p RegNo for pre-RA
|
|
|
|
/// and for post-RA, we give an assertion as without reaching definition
|
|
|
|
/// analysis post-RA, \p StartMI and \p EndMI are hard to keep right.
|
2020-08-10 22:55:22 +08:00
|
|
|
void fixupIsDeadOrKill(MachineInstr *StartMI, MachineInstr *EndMI,
|
2019-03-05 12:56:54 +08:00
|
|
|
unsigned RegNo) const;
|
2017-12-15 15:27:53 +08:00
|
|
|
void replaceInstrWithLI(MachineInstr &MI, const LoadImmediateInfo &LII) const;
|
2018-12-28 11:38:09 +08:00
|
|
|
void replaceInstrOperandWithImm(MachineInstr &MI, unsigned OpNo,
|
|
|
|
int64_t Imm) const;
|
2017-12-15 15:27:53 +08:00
|
|
|
|
2019-07-24 12:50:23 +08:00
|
|
|
bool instrHasImmForm(unsigned Opc, bool IsVFReg, ImmInstrInfo &III,
|
2018-10-22 19:22:59 +08:00
|
|
|
bool PostRA) const;
|
2018-09-27 19:49:47 +08:00
|
|
|
|
2019-07-25 16:36:44 +08:00
|
|
|
// In PostRA phase, try to find instruction defines \p Reg before \p MI.
|
|
|
|
// \p SeenIntermediate is set to true if uses between DefMI and \p MI exist.
|
|
|
|
MachineInstr *getDefMIPostRA(unsigned Reg, MachineInstr &MI,
|
|
|
|
bool &SeenIntermediateUse) const;
|
|
|
|
|
2018-09-27 19:49:47 +08:00
|
|
|
/// getRegNumForOperand - some operands use different numbering schemes
|
|
|
|
/// for the same registers. For example, a VSX instruction may have any of
|
|
|
|
/// vs0-vs63 allocated whereas an Altivec instruction could only have
|
|
|
|
/// vs32-vs63 allocated (numbered as v0-v31). This function returns the actual
|
|
|
|
/// register number needed for the opcode/operand number combination.
|
|
|
|
/// The operand number argument will be useful when we need to extend this
|
|
|
|
/// to instructions that use both Altivec and VSX numbering (for different
|
|
|
|
/// operands).
|
|
|
|
static unsigned getRegNumForOperand(const MCInstrDesc &Desc, unsigned Reg,
|
|
|
|
unsigned OpNo) {
|
2019-03-27 04:28:21 +08:00
|
|
|
int16_t regClass = Desc.OpInfo[OpNo].RegClass;
|
|
|
|
switch (regClass) {
|
|
|
|
// We store F0-F31, VF0-VF31 in MCOperand and it should be F0-F31,
|
|
|
|
// VSX32-VSX63 during encoding/disassembling
|
|
|
|
case PPC::VSSRCRegClassID:
|
|
|
|
case PPC::VSFRCRegClassID:
|
|
|
|
if (isVFRegister(Reg))
|
|
|
|
return PPC::VSX32 + (Reg - PPC::VF0);
|
|
|
|
break;
|
|
|
|
// We store VSL0-VSL31, V0-V31 in MCOperand and it should be VSL0-VSL31,
|
|
|
|
// VSX32-VSX63 during encoding/disassembling
|
|
|
|
case PPC::VSRCRegClassID:
|
|
|
|
if (isVRRegister(Reg))
|
|
|
|
return PPC::VSX32 + (Reg - PPC::V0);
|
|
|
|
break;
|
|
|
|
// Other RegClass doesn't need mapping
|
|
|
|
default:
|
|
|
|
break;
|
2018-09-27 19:49:47 +08:00
|
|
|
}
|
|
|
|
return Reg;
|
|
|
|
}
|
2019-06-12 01:40:39 +08:00
|
|
|
|
|
|
|
/// Check \p Opcode is BDNZ (Decrement CTR and branch if it is still nonzero).
|
|
|
|
bool isBDNZ(unsigned Opcode) const;
|
|
|
|
|
|
|
|
/// Find the hardware loop instruction used to set-up the specified loop.
|
|
|
|
/// On PPC, we have two instructions used to set-up the hardware loop
|
|
|
|
/// (MTCTRloop, MTCTR8loop) with corresponding endloop (BDNZ, BDNZ8)
|
|
|
|
/// instructions to indicate the end of a loop.
|
[MachinePipeliner] Improve the TargetInstrInfo API analyzeLoop/reduceLoopCount
Recommit: fix asan errors.
The way MachinePipeliner uses these target hooks is stateful - we reduce trip
count by one per call to reduceLoopCount. It's a little overfit for hardware
loops, where we don't have to worry about stitching a loop induction variable
across prologs and epilogs (the induction variable is implicit).
This patch introduces a new API:
/// Analyze loop L, which must be a single-basic-block loop, and if the
/// conditions can be understood enough produce a PipelinerLoopInfo object.
virtual std::unique_ptr<PipelinerLoopInfo>
analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const;
The return value is expected to be an implementation of the abstract class:
/// Object returned by analyzeLoopForPipelining. Allows software pipelining
/// implementations to query attributes of the loop being pipelined.
class PipelinerLoopInfo {
public:
virtual ~PipelinerLoopInfo();
/// Return true if the given instruction should not be pipelined and should
/// be ignored. An example could be a loop comparison, or induction variable
/// update with no users being pipelined.
virtual bool shouldIgnoreForPipelining(const MachineInstr *MI) const = 0;
/// Create a condition to determine if the trip count of the loop is greater
/// than TC.
///
/// If the trip count is statically known to be greater than TC, return
/// true. If the trip count is statically known to be not greater than TC,
/// return false. Otherwise return nullopt and fill out Cond with the test
/// condition.
virtual Optional<bool>
createTripCountGreaterCondition(int TC, MachineBasicBlock &MBB,
SmallVectorImpl<MachineOperand> &Cond) = 0;
/// Modify the loop such that the trip count is
/// OriginalTC + TripCountAdjust.
virtual void adjustTripCount(int TripCountAdjust) = 0;
/// Called when the loop's preheader has been modified to NewPreheader.
virtual void setPreheader(MachineBasicBlock *NewPreheader) = 0;
/// Called when the loop is being removed.
virtual void disposed() = 0;
};
The Pipeliner (ModuloSchedule.cpp) can use this object to modify the loop while
allowing the target to hold its own state across all calls. This API, in
particular the disjunction of creating a trip count check condition and
adjusting the loop, improves the code quality in ModuloSchedule.cpp.
llvm-svn: 372463
2019-09-21 16:19:41 +08:00
|
|
|
MachineInstr *
|
|
|
|
findLoopInstr(MachineBasicBlock &PreHeader,
|
|
|
|
SmallPtrSet<MachineBasicBlock *, 8> &Visited) const;
|
|
|
|
|
|
|
|
/// Analyze loop L, which must be a single-basic-block loop, and if the
|
|
|
|
/// conditions can be understood enough produce a PipelinerLoopInfo object.
|
|
|
|
std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
|
|
|
|
analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override;
|
2004-08-17 12:55:41 +08:00
|
|
|
};
|
|
|
|
|
2015-06-23 17:49:53 +08:00
|
|
|
}
|
2004-08-17 12:55:41 +08:00
|
|
|
|
|
|
|
#endif
|