2013-12-20 00:13:01 +08:00
|
|
|
//===------ PPCDisassembler.cpp - Disassembler for PowerPC ------*- C++ -*-===//
|
|
|
|
//
|
|
|
|
// The LLVM Compiler Infrastructure
|
|
|
|
//
|
|
|
|
// This file is distributed under the University of Illinois Open Source
|
|
|
|
// License. See LICENSE.TXT for details.
|
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
|
|
|
#include "PPC.h"
|
2016-01-27 00:44:37 +08:00
|
|
|
#include "llvm/MC/MCDisassembler/MCDisassembler.h"
|
2013-12-20 00:13:01 +08:00
|
|
|
#include "llvm/MC/MCFixedLenDisassembler.h"
|
|
|
|
#include "llvm/MC/MCInst.h"
|
|
|
|
#include "llvm/MC/MCSubtargetInfo.h"
|
2015-07-15 20:56:19 +08:00
|
|
|
#include "llvm/Support/Endian.h"
|
2013-12-20 00:13:01 +08:00
|
|
|
#include "llvm/Support/TargetRegistry.h"
|
|
|
|
|
|
|
|
using namespace llvm;
|
|
|
|
|
2014-04-22 06:55:11 +08:00
|
|
|
#define DEBUG_TYPE "ppc-disassembler"
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
typedef MCDisassembler::DecodeStatus DecodeStatus;
|
|
|
|
|
|
|
|
namespace {
|
|
|
|
class PPCDisassembler : public MCDisassembler {
|
2015-07-15 20:56:19 +08:00
|
|
|
bool IsLittleEndian;
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
public:
|
2015-07-15 20:56:19 +08:00
|
|
|
PPCDisassembler(const MCSubtargetInfo &STI, MCContext &Ctx,
|
|
|
|
bool IsLittleEndian)
|
|
|
|
: MCDisassembler(STI, Ctx), IsLittleEndian(IsLittleEndian) {}
|
2013-12-20 00:13:01 +08:00
|
|
|
|
2014-11-11 02:11:10 +08:00
|
|
|
DecodeStatus getInstruction(MCInst &Instr, uint64_t &Size,
|
2014-11-12 10:04:27 +08:00
|
|
|
ArrayRef<uint8_t> Bytes, uint64_t Address,
|
2014-11-11 02:11:10 +08:00
|
|
|
raw_ostream &VStream,
|
|
|
|
raw_ostream &CStream) const override;
|
2013-12-20 00:13:01 +08:00
|
|
|
};
|
|
|
|
} // end anonymous namespace
|
|
|
|
|
|
|
|
static MCDisassembler *createPPCDisassembler(const Target &T,
|
2014-04-15 12:40:56 +08:00
|
|
|
const MCSubtargetInfo &STI,
|
|
|
|
MCContext &Ctx) {
|
2015-07-15 20:56:19 +08:00
|
|
|
return new PPCDisassembler(STI, Ctx, /*IsLittleEndian=*/false);
|
|
|
|
}
|
|
|
|
|
|
|
|
static MCDisassembler *createPPCLEDisassembler(const Target &T,
|
|
|
|
const MCSubtargetInfo &STI,
|
|
|
|
MCContext &Ctx) {
|
|
|
|
return new PPCDisassembler(STI, Ctx, /*IsLittleEndian=*/true);
|
2013-12-20 00:13:01 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
extern "C" void LLVMInitializePowerPCDisassembler() {
|
|
|
|
// Register the disassembler for each target.
|
2016-10-10 07:00:34 +08:00
|
|
|
TargetRegistry::RegisterMCDisassembler(getThePPC32Target(),
|
2013-12-20 00:13:01 +08:00
|
|
|
createPPCDisassembler);
|
2016-10-10 07:00:34 +08:00
|
|
|
TargetRegistry::RegisterMCDisassembler(getThePPC64Target(),
|
2013-12-20 00:13:01 +08:00
|
|
|
createPPCDisassembler);
|
2016-10-10 07:00:34 +08:00
|
|
|
TargetRegistry::RegisterMCDisassembler(getThePPC64LETarget(),
|
2015-07-15 20:56:19 +08:00
|
|
|
createPPCLEDisassembler);
|
2013-12-20 00:13:01 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// FIXME: These can be generated by TableGen from the existing register
|
|
|
|
// encoding values!
|
|
|
|
|
|
|
|
static const unsigned CRRegs[] = {
|
|
|
|
PPC::CR0, PPC::CR1, PPC::CR2, PPC::CR3,
|
|
|
|
PPC::CR4, PPC::CR5, PPC::CR6, PPC::CR7
|
|
|
|
};
|
|
|
|
|
|
|
|
static const unsigned CRBITRegs[] = {
|
|
|
|
PPC::CR0LT, PPC::CR0GT, PPC::CR0EQ, PPC::CR0UN,
|
|
|
|
PPC::CR1LT, PPC::CR1GT, PPC::CR1EQ, PPC::CR1UN,
|
|
|
|
PPC::CR2LT, PPC::CR2GT, PPC::CR2EQ, PPC::CR2UN,
|
|
|
|
PPC::CR3LT, PPC::CR3GT, PPC::CR3EQ, PPC::CR3UN,
|
|
|
|
PPC::CR4LT, PPC::CR4GT, PPC::CR4EQ, PPC::CR4UN,
|
|
|
|
PPC::CR5LT, PPC::CR5GT, PPC::CR5EQ, PPC::CR5UN,
|
|
|
|
PPC::CR6LT, PPC::CR6GT, PPC::CR6EQ, PPC::CR6UN,
|
|
|
|
PPC::CR7LT, PPC::CR7GT, PPC::CR7EQ, PPC::CR7UN
|
|
|
|
};
|
|
|
|
|
|
|
|
static const unsigned FRegs[] = {
|
|
|
|
PPC::F0, PPC::F1, PPC::F2, PPC::F3,
|
|
|
|
PPC::F4, PPC::F5, PPC::F6, PPC::F7,
|
|
|
|
PPC::F8, PPC::F9, PPC::F10, PPC::F11,
|
|
|
|
PPC::F12, PPC::F13, PPC::F14, PPC::F15,
|
|
|
|
PPC::F16, PPC::F17, PPC::F18, PPC::F19,
|
|
|
|
PPC::F20, PPC::F21, PPC::F22, PPC::F23,
|
|
|
|
PPC::F24, PPC::F25, PPC::F26, PPC::F27,
|
|
|
|
PPC::F28, PPC::F29, PPC::F30, PPC::F31
|
|
|
|
};
|
|
|
|
|
2016-10-04 14:59:23 +08:00
|
|
|
static const unsigned VFRegs[] = {
|
|
|
|
PPC::VF0, PPC::VF1, PPC::VF2, PPC::VF3,
|
|
|
|
PPC::VF4, PPC::VF5, PPC::VF6, PPC::VF7,
|
|
|
|
PPC::VF8, PPC::VF9, PPC::VF10, PPC::VF11,
|
|
|
|
PPC::VF12, PPC::VF13, PPC::VF14, PPC::VF15,
|
|
|
|
PPC::VF16, PPC::VF17, PPC::VF18, PPC::VF19,
|
|
|
|
PPC::VF20, PPC::VF21, PPC::VF22, PPC::VF23,
|
|
|
|
PPC::VF24, PPC::VF25, PPC::VF26, PPC::VF27,
|
|
|
|
PPC::VF28, PPC::VF29, PPC::VF30, PPC::VF31
|
|
|
|
};
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
static const unsigned VRegs[] = {
|
|
|
|
PPC::V0, PPC::V1, PPC::V2, PPC::V3,
|
|
|
|
PPC::V4, PPC::V5, PPC::V6, PPC::V7,
|
|
|
|
PPC::V8, PPC::V9, PPC::V10, PPC::V11,
|
|
|
|
PPC::V12, PPC::V13, PPC::V14, PPC::V15,
|
|
|
|
PPC::V16, PPC::V17, PPC::V18, PPC::V19,
|
|
|
|
PPC::V20, PPC::V21, PPC::V22, PPC::V23,
|
|
|
|
PPC::V24, PPC::V25, PPC::V26, PPC::V27,
|
|
|
|
PPC::V28, PPC::V29, PPC::V30, PPC::V31
|
|
|
|
};
|
|
|
|
|
[PowerPC] Initial support for the VSX instruction set
VSX is an ISA extension supported on the POWER7 and later cores that enhances
floating-point vector and scalar capabilities. Among other things, this adds
<2 x double> support and generally helps to reduce register pressure.
The interesting part of this ISA feature is the register configuration: there
are 64 new 128-bit vector registers, the 32 of which are super-registers of the
existing 32 scalar floating-point registers, and the second 32 of which overlap
with the 32 Altivec vector registers. This makes things like vector insertion
and extraction tricky: this can be free but only if we force a restriction to
the right register subclass when needed. A new "minipass" PPCVSXCopy takes care
of this (although it could do a more-optimal job of it; see the comment about
unnecessary copies below).
Please note that, currently, VSX is not enabled by default when targeting
anything because it is not yet ready for that. The assembler and disassembler
are fully implemented and tested. However:
- CodeGen support causes miscompiles; test-suite runtime failures:
MultiSource/Benchmarks/FreeBench/distray/distray
MultiSource/Benchmarks/McCat/08-main/main
MultiSource/Benchmarks/Olden/voronoi/voronoi
MultiSource/Benchmarks/mafft/pairlocalalign
MultiSource/Benchmarks/tramp3d-v4/tramp3d-v4
SingleSource/Benchmarks/CoyoteBench/almabench
SingleSource/Benchmarks/Misc/matmul_f64_4x4
- The lowering currently falls back to using Altivec instructions far more
than it should. Worse, there are some things that are scalarized through the
stack that shouldn't be.
- A lot of unnecessary copies make it past the optimizers, and this needs to
be fixed.
- Many more regression tests are needed.
Normally, I'd fix these things prior to committing, but there are some
students and other contributors who would like to work this, and so it makes
sense to move this development process upstream where it can be subject to the
regular code-review procedures.
llvm-svn: 203768
2014-03-13 15:58:58 +08:00
|
|
|
static const unsigned VSRegs[] = {
|
|
|
|
PPC::VSL0, PPC::VSL1, PPC::VSL2, PPC::VSL3,
|
|
|
|
PPC::VSL4, PPC::VSL5, PPC::VSL6, PPC::VSL7,
|
|
|
|
PPC::VSL8, PPC::VSL9, PPC::VSL10, PPC::VSL11,
|
|
|
|
PPC::VSL12, PPC::VSL13, PPC::VSL14, PPC::VSL15,
|
|
|
|
PPC::VSL16, PPC::VSL17, PPC::VSL18, PPC::VSL19,
|
|
|
|
PPC::VSL20, PPC::VSL21, PPC::VSL22, PPC::VSL23,
|
|
|
|
PPC::VSL24, PPC::VSL25, PPC::VSL26, PPC::VSL27,
|
|
|
|
PPC::VSL28, PPC::VSL29, PPC::VSL30, PPC::VSL31,
|
|
|
|
|
2016-10-04 14:59:23 +08:00
|
|
|
PPC::V0, PPC::V1, PPC::V2, PPC::V3,
|
|
|
|
PPC::V4, PPC::V5, PPC::V6, PPC::V7,
|
|
|
|
PPC::V8, PPC::V9, PPC::V10, PPC::V11,
|
|
|
|
PPC::V12, PPC::V13, PPC::V14, PPC::V15,
|
|
|
|
PPC::V16, PPC::V17, PPC::V18, PPC::V19,
|
|
|
|
PPC::V20, PPC::V21, PPC::V22, PPC::V23,
|
|
|
|
PPC::V24, PPC::V25, PPC::V26, PPC::V27,
|
|
|
|
PPC::V28, PPC::V29, PPC::V30, PPC::V31
|
[PowerPC] Initial support for the VSX instruction set
VSX is an ISA extension supported on the POWER7 and later cores that enhances
floating-point vector and scalar capabilities. Among other things, this adds
<2 x double> support and generally helps to reduce register pressure.
The interesting part of this ISA feature is the register configuration: there
are 64 new 128-bit vector registers, the 32 of which are super-registers of the
existing 32 scalar floating-point registers, and the second 32 of which overlap
with the 32 Altivec vector registers. This makes things like vector insertion
and extraction tricky: this can be free but only if we force a restriction to
the right register subclass when needed. A new "minipass" PPCVSXCopy takes care
of this (although it could do a more-optimal job of it; see the comment about
unnecessary copies below).
Please note that, currently, VSX is not enabled by default when targeting
anything because it is not yet ready for that. The assembler and disassembler
are fully implemented and tested. However:
- CodeGen support causes miscompiles; test-suite runtime failures:
MultiSource/Benchmarks/FreeBench/distray/distray
MultiSource/Benchmarks/McCat/08-main/main
MultiSource/Benchmarks/Olden/voronoi/voronoi
MultiSource/Benchmarks/mafft/pairlocalalign
MultiSource/Benchmarks/tramp3d-v4/tramp3d-v4
SingleSource/Benchmarks/CoyoteBench/almabench
SingleSource/Benchmarks/Misc/matmul_f64_4x4
- The lowering currently falls back to using Altivec instructions far more
than it should. Worse, there are some things that are scalarized through the
stack that shouldn't be.
- A lot of unnecessary copies make it past the optimizers, and this needs to
be fixed.
- Many more regression tests are needed.
Normally, I'd fix these things prior to committing, but there are some
students and other contributors who would like to work this, and so it makes
sense to move this development process upstream where it can be subject to the
regular code-review procedures.
llvm-svn: 203768
2014-03-13 15:58:58 +08:00
|
|
|
};
|
|
|
|
|
2014-03-29 13:29:01 +08:00
|
|
|
static const unsigned VSFRegs[] = {
|
|
|
|
PPC::F0, PPC::F1, PPC::F2, PPC::F3,
|
|
|
|
PPC::F4, PPC::F5, PPC::F6, PPC::F7,
|
|
|
|
PPC::F8, PPC::F9, PPC::F10, PPC::F11,
|
|
|
|
PPC::F12, PPC::F13, PPC::F14, PPC::F15,
|
|
|
|
PPC::F16, PPC::F17, PPC::F18, PPC::F19,
|
|
|
|
PPC::F20, PPC::F21, PPC::F22, PPC::F23,
|
|
|
|
PPC::F24, PPC::F25, PPC::F26, PPC::F27,
|
|
|
|
PPC::F28, PPC::F29, PPC::F30, PPC::F31,
|
|
|
|
|
|
|
|
PPC::VF0, PPC::VF1, PPC::VF2, PPC::VF3,
|
|
|
|
PPC::VF4, PPC::VF5, PPC::VF6, PPC::VF7,
|
|
|
|
PPC::VF8, PPC::VF9, PPC::VF10, PPC::VF11,
|
|
|
|
PPC::VF12, PPC::VF13, PPC::VF14, PPC::VF15,
|
|
|
|
PPC::VF16, PPC::VF17, PPC::VF18, PPC::VF19,
|
|
|
|
PPC::VF20, PPC::VF21, PPC::VF22, PPC::VF23,
|
|
|
|
PPC::VF24, PPC::VF25, PPC::VF26, PPC::VF27,
|
|
|
|
PPC::VF28, PPC::VF29, PPC::VF30, PPC::VF31
|
|
|
|
};
|
|
|
|
|
2015-05-08 02:24:05 +08:00
|
|
|
static const unsigned VSSRegs[] = {
|
|
|
|
PPC::F0, PPC::F1, PPC::F2, PPC::F3,
|
|
|
|
PPC::F4, PPC::F5, PPC::F6, PPC::F7,
|
|
|
|
PPC::F8, PPC::F9, PPC::F10, PPC::F11,
|
|
|
|
PPC::F12, PPC::F13, PPC::F14, PPC::F15,
|
|
|
|
PPC::F16, PPC::F17, PPC::F18, PPC::F19,
|
|
|
|
PPC::F20, PPC::F21, PPC::F22, PPC::F23,
|
|
|
|
PPC::F24, PPC::F25, PPC::F26, PPC::F27,
|
|
|
|
PPC::F28, PPC::F29, PPC::F30, PPC::F31,
|
|
|
|
|
|
|
|
PPC::VF0, PPC::VF1, PPC::VF2, PPC::VF3,
|
|
|
|
PPC::VF4, PPC::VF5, PPC::VF6, PPC::VF7,
|
|
|
|
PPC::VF8, PPC::VF9, PPC::VF10, PPC::VF11,
|
|
|
|
PPC::VF12, PPC::VF13, PPC::VF14, PPC::VF15,
|
|
|
|
PPC::VF16, PPC::VF17, PPC::VF18, PPC::VF19,
|
|
|
|
PPC::VF20, PPC::VF21, PPC::VF22, PPC::VF23,
|
|
|
|
PPC::VF24, PPC::VF25, PPC::VF26, PPC::VF27,
|
|
|
|
PPC::VF28, PPC::VF29, PPC::VF30, PPC::VF31
|
|
|
|
};
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
static const unsigned GPRegs[] = {
|
|
|
|
PPC::R0, PPC::R1, PPC::R2, PPC::R3,
|
|
|
|
PPC::R4, PPC::R5, PPC::R6, PPC::R7,
|
|
|
|
PPC::R8, PPC::R9, PPC::R10, PPC::R11,
|
|
|
|
PPC::R12, PPC::R13, PPC::R14, PPC::R15,
|
|
|
|
PPC::R16, PPC::R17, PPC::R18, PPC::R19,
|
|
|
|
PPC::R20, PPC::R21, PPC::R22, PPC::R23,
|
|
|
|
PPC::R24, PPC::R25, PPC::R26, PPC::R27,
|
|
|
|
PPC::R28, PPC::R29, PPC::R30, PPC::R31
|
|
|
|
};
|
|
|
|
|
|
|
|
static const unsigned GP0Regs[] = {
|
|
|
|
PPC::ZERO, PPC::R1, PPC::R2, PPC::R3,
|
|
|
|
PPC::R4, PPC::R5, PPC::R6, PPC::R7,
|
|
|
|
PPC::R8, PPC::R9, PPC::R10, PPC::R11,
|
|
|
|
PPC::R12, PPC::R13, PPC::R14, PPC::R15,
|
|
|
|
PPC::R16, PPC::R17, PPC::R18, PPC::R19,
|
|
|
|
PPC::R20, PPC::R21, PPC::R22, PPC::R23,
|
|
|
|
PPC::R24, PPC::R25, PPC::R26, PPC::R27,
|
|
|
|
PPC::R28, PPC::R29, PPC::R30, PPC::R31
|
|
|
|
};
|
|
|
|
|
|
|
|
static const unsigned G8Regs[] = {
|
|
|
|
PPC::X0, PPC::X1, PPC::X2, PPC::X3,
|
|
|
|
PPC::X4, PPC::X5, PPC::X6, PPC::X7,
|
|
|
|
PPC::X8, PPC::X9, PPC::X10, PPC::X11,
|
|
|
|
PPC::X12, PPC::X13, PPC::X14, PPC::X15,
|
|
|
|
PPC::X16, PPC::X17, PPC::X18, PPC::X19,
|
|
|
|
PPC::X20, PPC::X21, PPC::X22, PPC::X23,
|
|
|
|
PPC::X24, PPC::X25, PPC::X26, PPC::X27,
|
|
|
|
PPC::X28, PPC::X29, PPC::X30, PPC::X31
|
|
|
|
};
|
|
|
|
|
2017-05-12 06:17:35 +08:00
|
|
|
static const unsigned G80Regs[] = {
|
|
|
|
PPC::ZERO8, PPC::X1, PPC::X2, PPC::X3,
|
|
|
|
PPC::X4, PPC::X5, PPC::X6, PPC::X7,
|
|
|
|
PPC::X8, PPC::X9, PPC::X10, PPC::X11,
|
|
|
|
PPC::X12, PPC::X13, PPC::X14, PPC::X15,
|
|
|
|
PPC::X16, PPC::X17, PPC::X18, PPC::X19,
|
|
|
|
PPC::X20, PPC::X21, PPC::X22, PPC::X23,
|
|
|
|
PPC::X24, PPC::X25, PPC::X26, PPC::X27,
|
|
|
|
PPC::X28, PPC::X29, PPC::X30, PPC::X31
|
|
|
|
};
|
|
|
|
|
[PowerPC] Add support for the QPX vector instruction set
This adds support for the QPX vector instruction set, which is used by the
enhanced A2 cores on the IBM BG/Q supercomputers. QPX vectors are 256 bytes
wide, holding 4 double-precision floating-point values. Boolean values, modeled
here as <4 x i1> are actually also represented as floating-point values
(essentially { -1, 1 } for { false, true }). QPX shares many features with
Altivec and VSX, but is distinct from both of them. One major difference is
that, instead of adding completely-separate vector registers, QPX vector
registers are extensions of the scalar floating-point registers (lane 0 is the
corresponding scalar floating-point value). The operations supported on QPX
vectors mirrors that supported on the scalar floating-point values (with some
additional ones for permutations and logical/comparison operations).
I've been maintaining this support out-of-tree, as part of the bgclang project,
for several years. This is not the entire bgclang patch set, but is most of the
subset that can be cleanly integrated into LLVM proper at this time. Adding
this to the LLVM backend is part of my efforts to rebase bgclang to the current
LLVM trunk, but is independently useful (especially for codes that use LLVM as
a JIT in library form).
The assembler/disassembler test coverage is complete. The CodeGen test coverage
is not, but I've included some tests, and more will be added as follow-up work.
llvm-svn: 230413
2015-02-25 09:06:45 +08:00
|
|
|
static const unsigned QFRegs[] = {
|
|
|
|
PPC::QF0, PPC::QF1, PPC::QF2, PPC::QF3,
|
|
|
|
PPC::QF4, PPC::QF5, PPC::QF6, PPC::QF7,
|
|
|
|
PPC::QF8, PPC::QF9, PPC::QF10, PPC::QF11,
|
|
|
|
PPC::QF12, PPC::QF13, PPC::QF14, PPC::QF15,
|
|
|
|
PPC::QF16, PPC::QF17, PPC::QF18, PPC::QF19,
|
|
|
|
PPC::QF20, PPC::QF21, PPC::QF22, PPC::QF23,
|
|
|
|
PPC::QF24, PPC::QF25, PPC::QF26, PPC::QF27,
|
|
|
|
PPC::QF28, PPC::QF29, PPC::QF30, PPC::QF31
|
|
|
|
};
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
template <std::size_t N>
|
|
|
|
static DecodeStatus decodeRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
const unsigned (&Regs)[N]) {
|
|
|
|
assert(RegNo < N && "Invalid register number");
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(Regs[RegNo]));
|
2013-12-20 00:13:01 +08:00
|
|
|
return MCDisassembler::Success;
|
|
|
|
}
|
|
|
|
|
|
|
|
static DecodeStatus DecodeCRRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, CRRegs);
|
|
|
|
}
|
|
|
|
|
2015-03-26 03:36:23 +08:00
|
|
|
static DecodeStatus DecodeCRRC0RegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, CRRegs);
|
|
|
|
}
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
static DecodeStatus DecodeCRBITRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, CRBITRegs);
|
|
|
|
}
|
|
|
|
|
|
|
|
static DecodeStatus DecodeF4RCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, FRegs);
|
|
|
|
}
|
|
|
|
|
|
|
|
static DecodeStatus DecodeF8RCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, FRegs);
|
|
|
|
}
|
|
|
|
|
2016-10-04 14:59:23 +08:00
|
|
|
static DecodeStatus DecodeVFRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, VFRegs);
|
|
|
|
}
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
static DecodeStatus DecodeVRRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, VRegs);
|
|
|
|
}
|
|
|
|
|
[PowerPC] Initial support for the VSX instruction set
VSX is an ISA extension supported on the POWER7 and later cores that enhances
floating-point vector and scalar capabilities. Among other things, this adds
<2 x double> support and generally helps to reduce register pressure.
The interesting part of this ISA feature is the register configuration: there
are 64 new 128-bit vector registers, the 32 of which are super-registers of the
existing 32 scalar floating-point registers, and the second 32 of which overlap
with the 32 Altivec vector registers. This makes things like vector insertion
and extraction tricky: this can be free but only if we force a restriction to
the right register subclass when needed. A new "minipass" PPCVSXCopy takes care
of this (although it could do a more-optimal job of it; see the comment about
unnecessary copies below).
Please note that, currently, VSX is not enabled by default when targeting
anything because it is not yet ready for that. The assembler and disassembler
are fully implemented and tested. However:
- CodeGen support causes miscompiles; test-suite runtime failures:
MultiSource/Benchmarks/FreeBench/distray/distray
MultiSource/Benchmarks/McCat/08-main/main
MultiSource/Benchmarks/Olden/voronoi/voronoi
MultiSource/Benchmarks/mafft/pairlocalalign
MultiSource/Benchmarks/tramp3d-v4/tramp3d-v4
SingleSource/Benchmarks/CoyoteBench/almabench
SingleSource/Benchmarks/Misc/matmul_f64_4x4
- The lowering currently falls back to using Altivec instructions far more
than it should. Worse, there are some things that are scalarized through the
stack that shouldn't be.
- A lot of unnecessary copies make it past the optimizers, and this needs to
be fixed.
- Many more regression tests are needed.
Normally, I'd fix these things prior to committing, but there are some
students and other contributors who would like to work this, and so it makes
sense to move this development process upstream where it can be subject to the
regular code-review procedures.
llvm-svn: 203768
2014-03-13 15:58:58 +08:00
|
|
|
static DecodeStatus DecodeVSRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, VSRegs);
|
|
|
|
}
|
|
|
|
|
2014-03-29 13:29:01 +08:00
|
|
|
static DecodeStatus DecodeVSFRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, VSFRegs);
|
|
|
|
}
|
|
|
|
|
2015-05-08 02:24:05 +08:00
|
|
|
static DecodeStatus DecodeVSSRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, VSSRegs);
|
|
|
|
}
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
static DecodeStatus DecodeGPRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, GPRegs);
|
|
|
|
}
|
|
|
|
|
|
|
|
static DecodeStatus DecodeGPRC_NOR0RegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, GP0Regs);
|
|
|
|
}
|
|
|
|
|
|
|
|
static DecodeStatus DecodeG8RCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, G8Regs);
|
|
|
|
}
|
|
|
|
|
2017-05-12 06:17:35 +08:00
|
|
|
static DecodeStatus DecodeG8RC_NOX0RegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, G80Regs);
|
|
|
|
}
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
#define DecodePointerLikeRegClass0 DecodeGPRCRegisterClass
|
|
|
|
#define DecodePointerLikeRegClass1 DecodeGPRC_NOR0RegisterClass
|
|
|
|
|
[PowerPC] Add support for the QPX vector instruction set
This adds support for the QPX vector instruction set, which is used by the
enhanced A2 cores on the IBM BG/Q supercomputers. QPX vectors are 256 bytes
wide, holding 4 double-precision floating-point values. Boolean values, modeled
here as <4 x i1> are actually also represented as floating-point values
(essentially { -1, 1 } for { false, true }). QPX shares many features with
Altivec and VSX, but is distinct from both of them. One major difference is
that, instead of adding completely-separate vector registers, QPX vector
registers are extensions of the scalar floating-point registers (lane 0 is the
corresponding scalar floating-point value). The operations supported on QPX
vectors mirrors that supported on the scalar floating-point values (with some
additional ones for permutations and logical/comparison operations).
I've been maintaining this support out-of-tree, as part of the bgclang project,
for several years. This is not the entire bgclang patch set, but is most of the
subset that can be cleanly integrated into LLVM proper at this time. Adding
this to the LLVM backend is part of my efforts to rebase bgclang to the current
LLVM trunk, but is independently useful (especially for codes that use LLVM as
a JIT in library form).
The assembler/disassembler test coverage is complete. The CodeGen test coverage
is not, but I've included some tests, and more will be added as follow-up work.
llvm-svn: 230413
2015-02-25 09:06:45 +08:00
|
|
|
static DecodeStatus DecodeQFRCRegisterClass(MCInst &Inst, uint64_t RegNo,
|
|
|
|
uint64_t Address,
|
|
|
|
const void *Decoder) {
|
|
|
|
return decodeRegisterClass(Inst, RegNo, QFRegs);
|
|
|
|
}
|
|
|
|
|
|
|
|
#define DecodeQSRCRegisterClass DecodeQFRCRegisterClass
|
|
|
|
#define DecodeQBRCRegisterClass DecodeQFRCRegisterClass
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
template<unsigned N>
|
|
|
|
static DecodeStatus decodeUImmOperand(MCInst &Inst, uint64_t Imm,
|
|
|
|
int64_t Address, const void *Decoder) {
|
|
|
|
assert(isUInt<N>(Imm) && "Invalid immediate");
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.addOperand(MCOperand::createImm(Imm));
|
2013-12-20 00:13:01 +08:00
|
|
|
return MCDisassembler::Success;
|
|
|
|
}
|
|
|
|
|
|
|
|
template<unsigned N>
|
|
|
|
static DecodeStatus decodeSImmOperand(MCInst &Inst, uint64_t Imm,
|
|
|
|
int64_t Address, const void *Decoder) {
|
|
|
|
assert(isUInt<N>(Imm) && "Invalid immediate");
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.addOperand(MCOperand::createImm(SignExtend64<N>(Imm)));
|
2013-12-20 00:13:01 +08:00
|
|
|
return MCDisassembler::Success;
|
|
|
|
}
|
|
|
|
|
|
|
|
static DecodeStatus decodeMemRIOperands(MCInst &Inst, uint64_t Imm,
|
|
|
|
int64_t Address, const void *Decoder) {
|
|
|
|
// Decode the memri field (imm, reg), which has the low 16-bits as the
|
|
|
|
// displacement and the next 5 bits as the register #.
|
|
|
|
|
|
|
|
uint64_t Base = Imm >> 16;
|
|
|
|
uint64_t Disp = Imm & 0xFFFF;
|
|
|
|
|
|
|
|
assert(Base < 32 && "Invalid base register");
|
|
|
|
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default: break;
|
|
|
|
case PPC::LBZU:
|
|
|
|
case PPC::LHAU:
|
|
|
|
case PPC::LHZU:
|
|
|
|
case PPC::LWZU:
|
|
|
|
case PPC::LFSU:
|
|
|
|
case PPC::LFDU:
|
|
|
|
// Add the tied output operand.
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(GP0Regs[Base]));
|
2013-12-20 00:13:01 +08:00
|
|
|
break;
|
|
|
|
case PPC::STBU:
|
|
|
|
case PPC::STHU:
|
|
|
|
case PPC::STWU:
|
|
|
|
case PPC::STFSU:
|
|
|
|
case PPC::STFDU:
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.insert(Inst.begin(), MCOperand::createReg(GP0Regs[Base]));
|
2013-12-20 00:13:01 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.addOperand(MCOperand::createImm(SignExtend64<16>(Disp)));
|
|
|
|
Inst.addOperand(MCOperand::createReg(GP0Regs[Base]));
|
2013-12-20 00:13:01 +08:00
|
|
|
return MCDisassembler::Success;
|
|
|
|
}
|
|
|
|
|
|
|
|
static DecodeStatus decodeMemRIXOperands(MCInst &Inst, uint64_t Imm,
|
|
|
|
int64_t Address, const void *Decoder) {
|
|
|
|
// Decode the memrix field (imm, reg), which has the low 14-bits as the
|
|
|
|
// displacement and the next 5 bits as the register #.
|
|
|
|
|
|
|
|
uint64_t Base = Imm >> 14;
|
|
|
|
uint64_t Disp = Imm & 0x3FFF;
|
|
|
|
|
|
|
|
assert(Base < 32 && "Invalid base register");
|
|
|
|
|
|
|
|
if (Inst.getOpcode() == PPC::LDU)
|
|
|
|
// Add the tied output operand.
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(GP0Regs[Base]));
|
2013-12-20 00:13:01 +08:00
|
|
|
else if (Inst.getOpcode() == PPC::STDU)
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.insert(Inst.begin(), MCOperand::createReg(GP0Regs[Base]));
|
2013-12-20 00:13:01 +08:00
|
|
|
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.addOperand(MCOperand::createImm(SignExtend64<16>(Disp << 2)));
|
|
|
|
Inst.addOperand(MCOperand::createReg(GP0Regs[Base]));
|
2013-12-20 00:13:01 +08:00
|
|
|
return MCDisassembler::Success;
|
|
|
|
}
|
|
|
|
|
2016-03-08 11:49:13 +08:00
|
|
|
static DecodeStatus decodeMemRIX16Operands(MCInst &Inst, uint64_t Imm,
|
|
|
|
int64_t Address, const void *Decoder) {
|
|
|
|
// Decode the memrix16 field (imm, reg), which has the low 12-bits as the
|
|
|
|
// displacement with 16-byte aligned, and the next 5 bits as the register #.
|
|
|
|
|
|
|
|
uint64_t Base = Imm >> 12;
|
|
|
|
uint64_t Disp = Imm & 0xFFF;
|
|
|
|
|
|
|
|
assert(Base < 32 && "Invalid base register");
|
|
|
|
|
|
|
|
Inst.addOperand(MCOperand::createImm(SignExtend64<16>(Disp << 4)));
|
|
|
|
Inst.addOperand(MCOperand::createReg(GP0Regs[Base]));
|
|
|
|
return MCDisassembler::Success;
|
|
|
|
}
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
static DecodeStatus decodeCRBitMOperand(MCInst &Inst, uint64_t Imm,
|
|
|
|
int64_t Address, const void *Decoder) {
|
|
|
|
// The cr bit encoding is 0x80 >> cr_reg_num.
|
|
|
|
|
|
|
|
unsigned Zeros = countTrailingZeros(Imm);
|
|
|
|
assert(Zeros < 8 && "Invalid CR bit value");
|
|
|
|
|
2015-05-14 02:37:00 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(CRRegs[7 - Zeros]));
|
2013-12-20 00:13:01 +08:00
|
|
|
return MCDisassembler::Success;
|
|
|
|
}
|
|
|
|
|
|
|
|
#include "PPCGenDisassemblerTables.inc"
|
|
|
|
|
|
|
|
DecodeStatus PPCDisassembler::getInstruction(MCInst &MI, uint64_t &Size,
|
2014-11-12 10:04:27 +08:00
|
|
|
ArrayRef<uint8_t> Bytes,
|
2014-11-11 02:11:10 +08:00
|
|
|
uint64_t Address, raw_ostream &OS,
|
|
|
|
raw_ostream &CS) const {
|
2013-12-20 00:13:01 +08:00
|
|
|
// Get the four bytes of the instruction.
|
|
|
|
Size = 4;
|
2014-11-12 10:04:27 +08:00
|
|
|
if (Bytes.size() < 4) {
|
2013-12-20 00:13:01 +08:00
|
|
|
Size = 0;
|
|
|
|
return MCDisassembler::Fail;
|
|
|
|
}
|
|
|
|
|
2015-07-15 20:56:19 +08:00
|
|
|
// Read the instruction in the proper endianness.
|
|
|
|
uint32_t Inst = IsLittleEndian ? support::endian::read32le(Bytes.data())
|
|
|
|
: support::endian::read32be(Bytes.data());
|
2013-12-20 00:13:01 +08:00
|
|
|
|
2015-05-26 18:47:10 +08:00
|
|
|
if (STI.getFeatureBits()[PPC::FeatureQPX]) {
|
[PowerPC] Add support for the QPX vector instruction set
This adds support for the QPX vector instruction set, which is used by the
enhanced A2 cores on the IBM BG/Q supercomputers. QPX vectors are 256 bytes
wide, holding 4 double-precision floating-point values. Boolean values, modeled
here as <4 x i1> are actually also represented as floating-point values
(essentially { -1, 1 } for { false, true }). QPX shares many features with
Altivec and VSX, but is distinct from both of them. One major difference is
that, instead of adding completely-separate vector registers, QPX vector
registers are extensions of the scalar floating-point registers (lane 0 is the
corresponding scalar floating-point value). The operations supported on QPX
vectors mirrors that supported on the scalar floating-point values (with some
additional ones for permutations and logical/comparison operations).
I've been maintaining this support out-of-tree, as part of the bgclang project,
for several years. This is not the entire bgclang patch set, but is most of the
subset that can be cleanly integrated into LLVM proper at this time. Adding
this to the LLVM backend is part of my efforts to rebase bgclang to the current
LLVM trunk, but is independently useful (especially for codes that use LLVM as
a JIT in library form).
The assembler/disassembler test coverage is complete. The CodeGen test coverage
is not, but I've included some tests, and more will be added as follow-up work.
llvm-svn: 230413
2015-02-25 09:06:45 +08:00
|
|
|
DecodeStatus result =
|
|
|
|
decodeInstruction(DecoderTableQPX32, MI, Inst, Address, this, STI);
|
|
|
|
if (result != MCDisassembler::Fail)
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2013-12-20 00:13:01 +08:00
|
|
|
return decodeInstruction(DecoderTable32, MI, Inst, Address, this, STI);
|
|
|
|
}
|
|
|
|
|