2003-10-06 03:27:59 +08:00
|
|
|
//===- CodeEmitterGen.cpp - Code Emitter Generator ------------------------===//
|
2005-04-22 08:00:37 +08:00
|
|
|
//
|
2003-10-21 04:20:30 +08:00
|
|
|
// The LLVM Compiler Infrastructure
|
|
|
|
//
|
2007-12-30 04:37:13 +08:00
|
|
|
// This file is distributed under the University of Illinois Open Source
|
|
|
|
// License. See LICENSE.TXT for details.
|
2005-04-22 08:00:37 +08:00
|
|
|
//
|
2003-10-21 04:20:30 +08:00
|
|
|
//===----------------------------------------------------------------------===//
|
2003-10-06 03:27:59 +08:00
|
|
|
//
|
2004-08-05 06:07:54 +08:00
|
|
|
// CodeEmitterGen uses the descriptions of instructions and their fields to
|
|
|
|
// construct an automated code emitter: a function that, given a MachineInstr,
|
|
|
|
// returns the (currently, 32-bit unsigned) value of the instruction.
|
2003-10-06 03:27:59 +08:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2016-12-01 01:48:10 +08:00
|
|
|
#include "CodeGenInstruction.h"
|
2004-08-10 03:10:43 +08:00
|
|
|
#include "CodeGenTarget.h"
|
Check that emitted instructions meet their predicates on all targets except ARM, Mips, and X86.
Summary:
* ARM is omitted from this patch because this check appears to expose bugs in this target.
* Mips is omitted from this patch because this check either detects bugs or deliberate
emission of instructions that don't satisfy their predicates. One deliberate
use is the SYNC instruction where the version with an operand is correctly
defined as requiring MIPS32 while the version without an operand is defined
as an alias of 'SYNC 0' and requires MIPS2.
* X86 is omitted from this patch because it doesn't use the tablegen-erated
MCCodeEmitter infrastructure.
Patches for ARM and Mips will follow.
Depends on D25617
Reviewers: tstellarAMD, jmolloy
Subscribers: wdng, jmolloy, aemerson, rengolin, arsenm, jyknight, nemanjai, nhaehnle, tstellarAMD, llvm-commits
Differential Revision: https://reviews.llvm.org/D25618
llvm-svn: 287439
2016-11-19 21:05:44 +08:00
|
|
|
#include "SubtargetFeatureInfo.h"
|
2016-12-01 01:48:10 +08:00
|
|
|
#include "llvm/ADT/ArrayRef.h"
|
2006-07-14 05:02:53 +08:00
|
|
|
#include "llvm/ADT/StringExtras.h"
|
2016-12-01 01:48:10 +08:00
|
|
|
#include "llvm/Support/Casting.h"
|
|
|
|
#include "llvm/Support/raw_ostream.h"
|
2012-12-04 18:37:14 +08:00
|
|
|
#include "llvm/TableGen/Record.h"
|
2012-06-11 23:37:55 +08:00
|
|
|
#include "llvm/TableGen/TableGenBackend.h"
|
2016-12-01 01:48:10 +08:00
|
|
|
#include <cassert>
|
|
|
|
#include <cstdint>
|
2010-12-13 09:05:54 +08:00
|
|
|
#include <map>
|
2016-12-01 01:48:10 +08:00
|
|
|
#include <set>
|
2012-06-11 23:37:55 +08:00
|
|
|
#include <string>
|
2016-12-01 01:48:10 +08:00
|
|
|
#include <utility>
|
2012-06-11 23:37:55 +08:00
|
|
|
#include <vector>
|
2016-12-01 01:48:10 +08:00
|
|
|
|
2004-08-01 11:55:39 +08:00
|
|
|
using namespace llvm;
|
2003-11-12 06:41:34 +08:00
|
|
|
|
2012-06-11 23:37:55 +08:00
|
|
|
namespace {
|
|
|
|
|
|
|
|
class CodeEmitterGen {
|
|
|
|
RecordKeeper &Records;
|
2016-12-01 01:48:10 +08:00
|
|
|
|
2012-06-11 23:37:55 +08:00
|
|
|
public:
|
|
|
|
CodeEmitterGen(RecordKeeper &R) : Records(R) {}
|
|
|
|
|
|
|
|
void run(raw_ostream &o);
|
2016-12-01 01:48:10 +08:00
|
|
|
|
2012-06-11 23:37:55 +08:00
|
|
|
private:
|
|
|
|
int getVariableBit(const std::string &VarName, BitsInit *BI, int bit);
|
|
|
|
std::string getInstructionCase(Record *R, CodeGenTarget &Target);
|
|
|
|
void AddCodeToMergeInOperand(Record *R, BitsInit *BI,
|
|
|
|
const std::string &VarName,
|
|
|
|
unsigned &NumberedOp,
|
[TableGen] Optionally forbid overlap between named and positional operands
There are currently two schemes for mapping instruction operands to
instruction-format variables for generating the instruction encoders and
decoders for the assembler and disassembler respectively: a) to map by name and
b) to map by position.
In the long run, we'd like to remove the position-based scheme and use only
name-based mapping. Unfortunately, the name-based scheme currently cannot deal
with complex operands (those with suboperands), and so we currently must use
the position-based scheme for those. On the other hand, the position-based
scheme cannot deal with (register) variables that are split into multiple
ranges. An upcoming commit to the PowerPC backend (adding VSX support) will
require this capability. While we could teach the position-based scheme to
handle that, since we'd like to move away from the position-based mapping
generally, it seems silly to teach it new tricks now. What makes more sense is
to allow for partial transitioning: use the name-based mapping when possible,
and only use the position-based scheme when necessary.
Now the problem is that mixing the two sensibly was not possible: the
position-based mapping would map based on position, but would not skip those
variables that were mapped by name. Instead, the two sets of assignments would
overlap. However, I cannot currently change the current behavior, because there
are some backends that rely on it [I think mistakenly, but I'll send a message
to llvmdev about that]. So I've added a new TableGen bit variable:
noNamedPositionallyEncodedOperands, that can be used to cause the
position-based mapping to skip variables mapped by name.
llvm-svn: 203767
2014-03-13 15:57:54 +08:00
|
|
|
std::set<unsigned> &NamedOpIndices,
|
2012-06-11 23:37:55 +08:00
|
|
|
std::string &Case, CodeGenTarget &Target);
|
|
|
|
|
|
|
|
};
|
|
|
|
|
2006-07-14 06:17:08 +08:00
|
|
|
// If the VarBitInit at position 'bit' matches the specified variable then
|
|
|
|
// return the variable bit position. Otherwise return -1.
|
2009-12-16 04:21:44 +08:00
|
|
|
int CodeEmitterGen::getVariableBit(const std::string &VarName,
|
2011-07-30 06:43:06 +08:00
|
|
|
BitsInit *BI, int bit) {
|
2012-10-11 04:24:43 +08:00
|
|
|
if (VarBitInit *VBI = dyn_cast<VarBitInit>(BI->getBit(bit))) {
|
|
|
|
if (VarInit *VI = dyn_cast<VarInit>(VBI->getBitVar()))
|
2010-11-15 14:42:13 +08:00
|
|
|
if (VI->getName() == VarName)
|
|
|
|
return VBI->getBitNum();
|
2012-10-11 04:24:43 +08:00
|
|
|
} else if (VarInit *VI = dyn_cast<VarInit>(BI->getBit(bit))) {
|
2011-04-29 01:51:45 +08:00
|
|
|
if (VI->getName() == VarName)
|
|
|
|
return 0;
|
|
|
|
}
|
2010-10-08 00:56:28 +08:00
|
|
|
|
2006-07-14 05:02:53 +08:00
|
|
|
return -1;
|
2010-10-08 00:56:28 +08:00
|
|
|
}
|
2006-07-14 05:02:53 +08:00
|
|
|
|
2010-11-15 14:59:17 +08:00
|
|
|
void CodeEmitterGen::
|
2011-07-30 06:43:06 +08:00
|
|
|
AddCodeToMergeInOperand(Record *R, BitsInit *BI, const std::string &VarName,
|
2011-07-12 07:06:52 +08:00
|
|
|
unsigned &NumberedOp,
|
[TableGen] Optionally forbid overlap between named and positional operands
There are currently two schemes for mapping instruction operands to
instruction-format variables for generating the instruction encoders and
decoders for the assembler and disassembler respectively: a) to map by name and
b) to map by position.
In the long run, we'd like to remove the position-based scheme and use only
name-based mapping. Unfortunately, the name-based scheme currently cannot deal
with complex operands (those with suboperands), and so we currently must use
the position-based scheme for those. On the other hand, the position-based
scheme cannot deal with (register) variables that are split into multiple
ranges. An upcoming commit to the PowerPC backend (adding VSX support) will
require this capability. While we could teach the position-based scheme to
handle that, since we'd like to move away from the position-based mapping
generally, it seems silly to teach it new tricks now. What makes more sense is
to allow for partial transitioning: use the name-based mapping when possible,
and only use the position-based scheme when necessary.
Now the problem is that mixing the two sensibly was not possible: the
position-based mapping would map based on position, but would not skip those
variables that were mapped by name. Instead, the two sets of assignments would
overlap. However, I cannot currently change the current behavior, because there
are some backends that rely on it [I think mistakenly, but I'll send a message
to llvmdev about that]. So I've added a new TableGen bit variable:
noNamedPositionallyEncodedOperands, that can be used to cause the
position-based mapping to skip variables mapped by name.
llvm-svn: 203767
2014-03-13 15:57:54 +08:00
|
|
|
std::set<unsigned> &NamedOpIndices,
|
2010-11-15 14:59:17 +08:00
|
|
|
std::string &Case, CodeGenTarget &Target) {
|
|
|
|
CodeGenInstruction &CGI = Target.getInstruction(R);
|
|
|
|
|
2010-11-15 15:09:28 +08:00
|
|
|
// Determine if VarName actually contributes to the Inst encoding.
|
|
|
|
int bit = BI->getNumBits()-1;
|
|
|
|
|
|
|
|
// Scan for a bit that this contributed to.
|
|
|
|
for (; bit >= 0; ) {
|
|
|
|
if (getVariableBit(VarName, BI, bit) != -1)
|
|
|
|
break;
|
|
|
|
|
|
|
|
--bit;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we found no bits, ignore this value, otherwise emit the call to get the
|
|
|
|
// operand encoding.
|
|
|
|
if (bit < 0) return;
|
|
|
|
|
|
|
|
// If the operand matches by name, reference according to that
|
|
|
|
// operand number. Non-matching operands are assumed to be in
|
|
|
|
// order.
|
|
|
|
unsigned OpIdx;
|
|
|
|
if (CGI.Operands.hasOperandNamed(VarName, OpIdx)) {
|
|
|
|
// Get the machine operand number for the indicated operand.
|
|
|
|
OpIdx = CGI.Operands[OpIdx].MIOperandNo;
|
|
|
|
assert(!CGI.Operands.isFlatOperandNotEmitted(OpIdx) &&
|
|
|
|
"Explicitly used operand also marked as not emitted!");
|
|
|
|
} else {
|
2012-11-10 04:29:37 +08:00
|
|
|
unsigned NumberOps = CGI.Operands.size();
|
2010-11-15 15:09:28 +08:00
|
|
|
/// If this operand is not supposed to be emitted by the
|
|
|
|
/// generated emitter, skip it.
|
2012-11-10 04:29:37 +08:00
|
|
|
while (NumberedOp < NumberOps &&
|
[TableGen] Optionally forbid overlap between named and positional operands
There are currently two schemes for mapping instruction operands to
instruction-format variables for generating the instruction encoders and
decoders for the assembler and disassembler respectively: a) to map by name and
b) to map by position.
In the long run, we'd like to remove the position-based scheme and use only
name-based mapping. Unfortunately, the name-based scheme currently cannot deal
with complex operands (those with suboperands), and so we currently must use
the position-based scheme for those. On the other hand, the position-based
scheme cannot deal with (register) variables that are split into multiple
ranges. An upcoming commit to the PowerPC backend (adding VSX support) will
require this capability. While we could teach the position-based scheme to
handle that, since we'd like to move away from the position-based mapping
generally, it seems silly to teach it new tricks now. What makes more sense is
to allow for partial transitioning: use the name-based mapping when possible,
and only use the position-based scheme when necessary.
Now the problem is that mixing the two sensibly was not possible: the
position-based mapping would map based on position, but would not skip those
variables that were mapped by name. Instead, the two sets of assignments would
overlap. However, I cannot currently change the current behavior, because there
are some backends that rely on it [I think mistakenly, but I'll send a message
to llvmdev about that]. So I've added a new TableGen bit variable:
noNamedPositionallyEncodedOperands, that can be used to cause the
position-based mapping to skip variables mapped by name.
llvm-svn: 203767
2014-03-13 15:57:54 +08:00
|
|
|
(CGI.Operands.isFlatOperandNotEmitted(NumberedOp) ||
|
2015-01-15 19:41:30 +08:00
|
|
|
(!NamedOpIndices.empty() && NamedOpIndices.count(
|
2014-03-22 19:33:32 +08:00
|
|
|
CGI.Operands.getSubOperandNumber(NumberedOp).first)))) {
|
2010-11-15 15:09:28 +08:00
|
|
|
++NumberedOp;
|
2012-11-10 05:27:03 +08:00
|
|
|
|
2014-03-22 19:33:32 +08:00
|
|
|
if (NumberedOp >= CGI.Operands.back().MIOperandNo +
|
|
|
|
CGI.Operands.back().MINumOperands) {
|
|
|
|
errs() << "Too few operands in record " << R->getName() <<
|
|
|
|
" (no match for variable " << VarName << "):\n";
|
|
|
|
errs() << *R;
|
|
|
|
errs() << '\n';
|
|
|
|
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-11-15 15:09:28 +08:00
|
|
|
OpIdx = NumberedOp++;
|
|
|
|
}
|
|
|
|
|
|
|
|
std::pair<unsigned, unsigned> SO = CGI.Operands.getSubOperandNumber(OpIdx);
|
|
|
|
std::string &EncoderMethodName = CGI.Operands[SO.first].EncoderMethodName;
|
|
|
|
|
|
|
|
// If the source operand has a custom encoder, use it. This will
|
|
|
|
// get the encoding for all of the suboperands.
|
|
|
|
if (!EncoderMethodName.empty()) {
|
|
|
|
// A custom encoder has all of the information for the
|
|
|
|
// sub-operands, if there are more than one, so only
|
|
|
|
// query the encoder once per source operand.
|
|
|
|
if (SO.second == 0) {
|
|
|
|
Case += " // op: " + VarName + "\n" +
|
|
|
|
" op = " + EncoderMethodName + "(MI, " + utostr(OpIdx);
|
2014-09-03 06:28:02 +08:00
|
|
|
Case += ", Fixups, STI";
|
2010-11-15 15:09:28 +08:00
|
|
|
Case += ");\n";
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
Case += " // op: " + VarName + "\n" +
|
|
|
|
" op = getMachineOpValue(MI, MI.getOperand(" + utostr(OpIdx) + ")";
|
2014-09-03 06:28:02 +08:00
|
|
|
Case += ", Fixups, STI";
|
2010-11-15 15:09:28 +08:00
|
|
|
Case += ");\n";
|
|
|
|
}
|
|
|
|
|
|
|
|
for (; bit >= 0; ) {
|
2010-11-15 14:59:17 +08:00
|
|
|
int varBit = getVariableBit(VarName, BI, bit);
|
|
|
|
|
|
|
|
// If this bit isn't from a variable, skip it.
|
|
|
|
if (varBit == -1) {
|
|
|
|
--bit;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2011-01-28 07:08:52 +08:00
|
|
|
// Figure out the consecutive range of bits covered by this operand, in
|
2010-11-15 14:59:17 +08:00
|
|
|
// order to generate better encoding code.
|
|
|
|
int beginInstBit = bit;
|
|
|
|
int beginVarBit = varBit;
|
|
|
|
int N = 1;
|
|
|
|
for (--bit; bit >= 0;) {
|
|
|
|
varBit = getVariableBit(VarName, BI, bit);
|
|
|
|
if (varBit == -1 || varBit != (beginVarBit - N)) break;
|
|
|
|
++N;
|
|
|
|
--bit;
|
|
|
|
}
|
2010-11-15 15:09:28 +08:00
|
|
|
|
2012-03-09 22:52:44 +08:00
|
|
|
uint64_t opMask = ~(uint64_t)0 >> (64-N);
|
2010-11-15 14:59:17 +08:00
|
|
|
int opShift = beginVarBit - N + 1;
|
|
|
|
opMask <<= opShift;
|
|
|
|
opShift = beginInstBit - beginVarBit;
|
|
|
|
|
|
|
|
if (opShift > 0) {
|
2012-01-25 02:37:29 +08:00
|
|
|
Case += " Value |= (op & UINT64_C(" + utostr(opMask) + ")) << " +
|
2010-11-15 14:59:17 +08:00
|
|
|
itostr(opShift) + ";\n";
|
|
|
|
} else if (opShift < 0) {
|
2012-01-25 02:37:29 +08:00
|
|
|
Case += " Value |= (op & UINT64_C(" + utostr(opMask) + ")) >> " +
|
2010-11-15 14:59:17 +08:00
|
|
|
itostr(-opShift) + ";\n";
|
|
|
|
} else {
|
2012-01-25 02:37:29 +08:00
|
|
|
Case += " Value |= op & UINT64_C(" + utostr(opMask) + ");\n";
|
2010-11-15 14:59:17 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
std::string CodeEmitterGen::getInstructionCase(Record *R,
|
|
|
|
CodeGenTarget &Target) {
|
|
|
|
std::string Case;
|
2011-07-30 06:43:06 +08:00
|
|
|
BitsInit *BI = R->getValueAsBitsInit("Inst");
|
2010-11-15 14:59:17 +08:00
|
|
|
unsigned NumberedOp = 0;
|
[TableGen] Optionally forbid overlap between named and positional operands
There are currently two schemes for mapping instruction operands to
instruction-format variables for generating the instruction encoders and
decoders for the assembler and disassembler respectively: a) to map by name and
b) to map by position.
In the long run, we'd like to remove the position-based scheme and use only
name-based mapping. Unfortunately, the name-based scheme currently cannot deal
with complex operands (those with suboperands), and so we currently must use
the position-based scheme for those. On the other hand, the position-based
scheme cannot deal with (register) variables that are split into multiple
ranges. An upcoming commit to the PowerPC backend (adding VSX support) will
require this capability. While we could teach the position-based scheme to
handle that, since we'd like to move away from the position-based mapping
generally, it seems silly to teach it new tricks now. What makes more sense is
to allow for partial transitioning: use the name-based mapping when possible,
and only use the position-based scheme when necessary.
Now the problem is that mixing the two sensibly was not possible: the
position-based mapping would map based on position, but would not skip those
variables that were mapped by name. Instead, the two sets of assignments would
overlap. However, I cannot currently change the current behavior, because there
are some backends that rely on it [I think mistakenly, but I'll send a message
to llvmdev about that]. So I've added a new TableGen bit variable:
noNamedPositionallyEncodedOperands, that can be used to cause the
position-based mapping to skip variables mapped by name.
llvm-svn: 203767
2014-03-13 15:57:54 +08:00
|
|
|
std::set<unsigned> NamedOpIndices;
|
2017-07-04 14:16:53 +08:00
|
|
|
|
[TableGen] Optionally forbid overlap between named and positional operands
There are currently two schemes for mapping instruction operands to
instruction-format variables for generating the instruction encoders and
decoders for the assembler and disassembler respectively: a) to map by name and
b) to map by position.
In the long run, we'd like to remove the position-based scheme and use only
name-based mapping. Unfortunately, the name-based scheme currently cannot deal
with complex operands (those with suboperands), and so we currently must use
the position-based scheme for those. On the other hand, the position-based
scheme cannot deal with (register) variables that are split into multiple
ranges. An upcoming commit to the PowerPC backend (adding VSX support) will
require this capability. While we could teach the position-based scheme to
handle that, since we'd like to move away from the position-based mapping
generally, it seems silly to teach it new tricks now. What makes more sense is
to allow for partial transitioning: use the name-based mapping when possible,
and only use the position-based scheme when necessary.
Now the problem is that mixing the two sensibly was not possible: the
position-based mapping would map based on position, but would not skip those
variables that were mapped by name. Instead, the two sets of assignments would
overlap. However, I cannot currently change the current behavior, because there
are some backends that rely on it [I think mistakenly, but I'll send a message
to llvmdev about that]. So I've added a new TableGen bit variable:
noNamedPositionallyEncodedOperands, that can be used to cause the
position-based mapping to skip variables mapped by name.
llvm-svn: 203767
2014-03-13 15:57:54 +08:00
|
|
|
// Collect the set of operand indices that might correspond to named
|
|
|
|
// operand, and skip these when assigning operands based on position.
|
|
|
|
if (Target.getInstructionSet()->
|
|
|
|
getValueAsBit("noNamedPositionallyEncodedOperands")) {
|
|
|
|
CodeGenInstruction &CGI = Target.getInstruction(R);
|
2017-07-04 14:16:53 +08:00
|
|
|
for (const RecordVal &RV : R->getValues()) {
|
[TableGen] Optionally forbid overlap between named and positional operands
There are currently two schemes for mapping instruction operands to
instruction-format variables for generating the instruction encoders and
decoders for the assembler and disassembler respectively: a) to map by name and
b) to map by position.
In the long run, we'd like to remove the position-based scheme and use only
name-based mapping. Unfortunately, the name-based scheme currently cannot deal
with complex operands (those with suboperands), and so we currently must use
the position-based scheme for those. On the other hand, the position-based
scheme cannot deal with (register) variables that are split into multiple
ranges. An upcoming commit to the PowerPC backend (adding VSX support) will
require this capability. While we could teach the position-based scheme to
handle that, since we'd like to move away from the position-based mapping
generally, it seems silly to teach it new tricks now. What makes more sense is
to allow for partial transitioning: use the name-based mapping when possible,
and only use the position-based scheme when necessary.
Now the problem is that mixing the two sensibly was not possible: the
position-based mapping would map based on position, but would not skip those
variables that were mapped by name. Instead, the two sets of assignments would
overlap. However, I cannot currently change the current behavior, because there
are some backends that rely on it [I think mistakenly, but I'll send a message
to llvmdev about that]. So I've added a new TableGen bit variable:
noNamedPositionallyEncodedOperands, that can be used to cause the
position-based mapping to skip variables mapped by name.
llvm-svn: 203767
2014-03-13 15:57:54 +08:00
|
|
|
unsigned OpIdx;
|
2017-07-04 14:16:53 +08:00
|
|
|
if (!CGI.Operands.hasOperandNamed(RV.getName(), OpIdx))
|
[TableGen] Optionally forbid overlap between named and positional operands
There are currently two schemes for mapping instruction operands to
instruction-format variables for generating the instruction encoders and
decoders for the assembler and disassembler respectively: a) to map by name and
b) to map by position.
In the long run, we'd like to remove the position-based scheme and use only
name-based mapping. Unfortunately, the name-based scheme currently cannot deal
with complex operands (those with suboperands), and so we currently must use
the position-based scheme for those. On the other hand, the position-based
scheme cannot deal with (register) variables that are split into multiple
ranges. An upcoming commit to the PowerPC backend (adding VSX support) will
require this capability. While we could teach the position-based scheme to
handle that, since we'd like to move away from the position-based mapping
generally, it seems silly to teach it new tricks now. What makes more sense is
to allow for partial transitioning: use the name-based mapping when possible,
and only use the position-based scheme when necessary.
Now the problem is that mixing the two sensibly was not possible: the
position-based mapping would map based on position, but would not skip those
variables that were mapped by name. Instead, the two sets of assignments would
overlap. However, I cannot currently change the current behavior, because there
are some backends that rely on it [I think mistakenly, but I'll send a message
to llvmdev about that]. So I've added a new TableGen bit variable:
noNamedPositionallyEncodedOperands, that can be used to cause the
position-based mapping to skip variables mapped by name.
llvm-svn: 203767
2014-03-13 15:57:54 +08:00
|
|
|
continue;
|
|
|
|
|
|
|
|
NamedOpIndices.insert(OpIdx);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-11-15 14:59:17 +08:00
|
|
|
// Loop over all of the fields in the instruction, determining which are the
|
|
|
|
// operands to the instruction.
|
2017-07-04 14:16:53 +08:00
|
|
|
for (const RecordVal &RV : R->getValues()) {
|
2010-11-15 14:59:17 +08:00
|
|
|
// Ignore fixed fields in the record, we're looking for values like:
|
|
|
|
// bits<5> RST = { ?, ?, ?, ?, ? };
|
2017-07-04 14:16:53 +08:00
|
|
|
if (RV.getPrefix() || RV.getValue()->isComplete())
|
2010-11-15 14:59:17 +08:00
|
|
|
continue;
|
|
|
|
|
2017-07-04 14:16:53 +08:00
|
|
|
AddCodeToMergeInOperand(R, BI, RV.getName(), NumberedOp,
|
[TableGen] Optionally forbid overlap between named and positional operands
There are currently two schemes for mapping instruction operands to
instruction-format variables for generating the instruction encoders and
decoders for the assembler and disassembler respectively: a) to map by name and
b) to map by position.
In the long run, we'd like to remove the position-based scheme and use only
name-based mapping. Unfortunately, the name-based scheme currently cannot deal
with complex operands (those with suboperands), and so we currently must use
the position-based scheme for those. On the other hand, the position-based
scheme cannot deal with (register) variables that are split into multiple
ranges. An upcoming commit to the PowerPC backend (adding VSX support) will
require this capability. While we could teach the position-based scheme to
handle that, since we'd like to move away from the position-based mapping
generally, it seems silly to teach it new tricks now. What makes more sense is
to allow for partial transitioning: use the name-based mapping when possible,
and only use the position-based scheme when necessary.
Now the problem is that mixing the two sensibly was not possible: the
position-based mapping would map based on position, but would not skip those
variables that were mapped by name. Instead, the two sets of assignments would
overlap. However, I cannot currently change the current behavior, because there
are some backends that rely on it [I think mistakenly, but I'll send a message
to llvmdev about that]. So I've added a new TableGen bit variable:
noNamedPositionallyEncodedOperands, that can be used to cause the
position-based mapping to skip variables mapped by name.
llvm-svn: 203767
2014-03-13 15:57:54 +08:00
|
|
|
NamedOpIndices, Case, Target);
|
2010-11-15 14:59:17 +08:00
|
|
|
}
|
2017-06-01 05:12:46 +08:00
|
|
|
|
|
|
|
StringRef PostEmitter = R->getValueAsString("PostEncoderMethod");
|
2014-01-29 07:13:18 +08:00
|
|
|
if (!PostEmitter.empty()) {
|
2017-06-01 05:12:46 +08:00
|
|
|
Case += " Value = ";
|
|
|
|
Case += PostEmitter;
|
|
|
|
Case += "(MI, Value";
|
2014-09-03 06:28:02 +08:00
|
|
|
Case += ", STI";
|
2014-01-29 07:13:18 +08:00
|
|
|
Case += ");\n";
|
|
|
|
}
|
2010-11-15 14:59:17 +08:00
|
|
|
|
|
|
|
return Case;
|
|
|
|
}
|
|
|
|
|
2009-07-03 08:10:29 +08:00
|
|
|
void CodeEmitterGen::run(raw_ostream &o) {
|
2010-12-13 08:23:57 +08:00
|
|
|
CodeGenTarget Target(Records);
|
2003-10-06 03:27:59 +08:00
|
|
|
std::vector<Record*> Insts = Records.getAllDerivedDefinitions("Instruction");
|
2010-10-08 00:56:28 +08:00
|
|
|
|
2006-07-14 05:02:53 +08:00
|
|
|
// For little-endian instruction bit encodings, reverse the bit order
|
2013-12-18 06:37:50 +08:00
|
|
|
Target.reverseBitsForLittleEndianEncoding();
|
2010-10-08 00:56:28 +08:00
|
|
|
|
2016-02-01 09:33:42 +08:00
|
|
|
ArrayRef<const CodeGenInstruction*> NumberedInstructions =
|
2010-03-19 08:34:35 +08:00
|
|
|
Target.getInstructionsByEnumValue();
|
2003-10-06 03:27:59 +08:00
|
|
|
|
2004-08-11 04:54:58 +08:00
|
|
|
// Emit function declaration
|
2012-01-25 02:37:29 +08:00
|
|
|
o << "uint64_t " << Target.getName();
|
2014-09-03 06:28:02 +08:00
|
|
|
o << "MCCodeEmitter::getBinaryCodeForInstr(const MCInst &MI,\n"
|
|
|
|
<< " SmallVectorImpl<MCFixup> &Fixups,\n"
|
|
|
|
<< " const MCSubtargetInfo &STI) const {\n";
|
2004-08-11 04:54:58 +08:00
|
|
|
|
2006-07-13 03:15:43 +08:00
|
|
|
// Emit instruction base values
|
2012-03-07 05:48:32 +08:00
|
|
|
o << " static const uint64_t InstBits[] = {\n";
|
2016-02-01 09:33:42 +08:00
|
|
|
for (const CodeGenInstruction *CGI : NumberedInstructions) {
|
2006-07-13 03:15:43 +08:00
|
|
|
Record *R = CGI->TheDef;
|
2010-10-08 00:56:28 +08:00
|
|
|
|
2011-07-07 05:33:38 +08:00
|
|
|
if (R->getValueAsString("Namespace") == "TargetOpcode" ||
|
|
|
|
R->getValueAsBit("isPseudo")) {
|
2012-01-25 02:37:29 +08:00
|
|
|
o << " UINT64_C(0),\n";
|
2006-07-13 03:15:43 +08:00
|
|
|
continue;
|
|
|
|
}
|
2010-10-08 00:56:28 +08:00
|
|
|
|
2011-07-30 06:43:06 +08:00
|
|
|
BitsInit *BI = R->getValueAsBitsInit("Inst");
|
2003-10-06 03:27:59 +08:00
|
|
|
|
2010-11-15 14:59:17 +08:00
|
|
|
// Start by filling in fixed values.
|
2012-03-07 05:48:32 +08:00
|
|
|
uint64_t Value = 0;
|
2003-10-06 03:27:59 +08:00
|
|
|
for (unsigned i = 0, e = BI->getNumBits(); i != e; ++i) {
|
2012-10-11 04:24:43 +08:00
|
|
|
if (BitInit *B = dyn_cast<BitInit>(BI->getBit(e-i-1)))
|
2012-03-07 05:48:32 +08:00
|
|
|
Value |= (uint64_t)B->getValue() << (e-i-1);
|
2003-10-06 03:27:59 +08:00
|
|
|
}
|
2012-01-25 02:37:29 +08:00
|
|
|
o << " UINT64_C(" << Value << ")," << '\t' << "// " << R->getName() << "\n";
|
2006-07-13 03:15:43 +08:00
|
|
|
}
|
2012-01-25 02:37:29 +08:00
|
|
|
o << " UINT64_C(0)\n };\n";
|
2010-10-08 00:56:28 +08:00
|
|
|
|
2006-07-14 05:02:53 +08:00
|
|
|
// Map to accumulate all the cases.
|
2016-12-01 01:48:10 +08:00
|
|
|
std::map<std::string, std::vector<std::string>> CaseMap;
|
2010-10-08 00:56:28 +08:00
|
|
|
|
2006-07-14 05:02:53 +08:00
|
|
|
// Construct all cases statement for each opcode
|
|
|
|
for (std::vector<Record*>::iterator IC = Insts.begin(), EC = Insts.end();
|
|
|
|
IC != EC; ++IC) {
|
|
|
|
Record *R = *IC;
|
2011-07-07 05:33:38 +08:00
|
|
|
if (R->getValueAsString("Namespace") == "TargetOpcode" ||
|
|
|
|
R->getValueAsBit("isPseudo"))
|
2010-07-03 05:44:22 +08:00
|
|
|
continue;
|
2017-06-01 03:01:11 +08:00
|
|
|
std::string InstName =
|
|
|
|
(R->getValueAsString("Namespace") + "::" + R->getName()).str();
|
2010-11-15 14:59:17 +08:00
|
|
|
std::string Case = getInstructionCase(R, Target);
|
2010-11-11 09:19:24 +08:00
|
|
|
|
2017-06-01 03:01:11 +08:00
|
|
|
CaseMap[Case].push_back(std::move(InstName));
|
2006-07-14 05:02:53 +08:00
|
|
|
}
|
2005-04-22 08:00:37 +08:00
|
|
|
|
2006-07-14 05:02:53 +08:00
|
|
|
// Emit initial function code
|
|
|
|
o << " const unsigned opcode = MI.getOpcode();\n"
|
2012-03-07 05:48:32 +08:00
|
|
|
<< " uint64_t Value = InstBits[opcode];\n"
|
|
|
|
<< " uint64_t op = 0;\n"
|
2010-12-23 08:58:24 +08:00
|
|
|
<< " (void)op; // suppress warning\n"
|
2006-07-14 05:02:53 +08:00
|
|
|
<< " switch (opcode) {\n";
|
2005-04-22 08:00:37 +08:00
|
|
|
|
2006-07-14 05:02:53 +08:00
|
|
|
// Emit each case statement
|
2016-12-01 01:48:10 +08:00
|
|
|
std::map<std::string, std::vector<std::string>>::iterator IE, EE;
|
2006-07-14 05:02:53 +08:00
|
|
|
for (IE = CaseMap.begin(), EE = CaseMap.end(); IE != EE; ++IE) {
|
|
|
|
const std::string &Case = IE->first;
|
|
|
|
std::vector<std::string> &InstList = IE->second;
|
2005-04-22 08:00:37 +08:00
|
|
|
|
2006-07-14 05:02:53 +08:00
|
|
|
for (int i = 0, N = InstList.size(); i < N; i++) {
|
|
|
|
if (i) o << "\n";
|
2011-02-04 07:26:36 +08:00
|
|
|
o << " case " << InstList[i] << ":";
|
2003-10-06 03:27:59 +08:00
|
|
|
}
|
2006-07-14 05:02:53 +08:00
|
|
|
o << " {\n";
|
|
|
|
o << Case;
|
2003-10-06 03:27:59 +08:00
|
|
|
o << " break;\n"
|
|
|
|
<< " }\n";
|
|
|
|
}
|
|
|
|
|
2004-10-14 13:53:01 +08:00
|
|
|
// Default case: unhandled opcode
|
2003-10-06 03:27:59 +08:00
|
|
|
o << " default:\n"
|
2014-06-27 06:52:05 +08:00
|
|
|
<< " std::string msg;\n"
|
|
|
|
<< " raw_string_ostream Msg(msg);\n"
|
2009-07-09 03:04:27 +08:00
|
|
|
<< " Msg << \"Not supported instr: \" << MI;\n"
|
2010-04-08 06:58:41 +08:00
|
|
|
<< " report_fatal_error(Msg.str());\n"
|
2003-10-06 03:27:59 +08:00
|
|
|
<< " }\n"
|
|
|
|
<< " return Value;\n"
|
2004-10-14 13:53:01 +08:00
|
|
|
<< "}\n\n";
|
Check that emitted instructions meet their predicates on all targets except ARM, Mips, and X86.
Summary:
* ARM is omitted from this patch because this check appears to expose bugs in this target.
* Mips is omitted from this patch because this check either detects bugs or deliberate
emission of instructions that don't satisfy their predicates. One deliberate
use is the SYNC instruction where the version with an operand is correctly
defined as requiring MIPS32 while the version without an operand is defined
as an alias of 'SYNC 0' and requires MIPS2.
* X86 is omitted from this patch because it doesn't use the tablegen-erated
MCCodeEmitter infrastructure.
Patches for ARM and Mips will follow.
Depends on D25617
Reviewers: tstellarAMD, jmolloy
Subscribers: wdng, jmolloy, aemerson, rengolin, arsenm, jyknight, nemanjai, nhaehnle, tstellarAMD, llvm-commits
Differential Revision: https://reviews.llvm.org/D25618
llvm-svn: 287439
2016-11-19 21:05:44 +08:00
|
|
|
|
|
|
|
const auto &All = SubtargetFeatureInfo::getAll(Records);
|
|
|
|
std::map<Record *, SubtargetFeatureInfo, LessRecordByID> SubtargetFeatures;
|
|
|
|
SubtargetFeatures.insert(All.begin(), All.end());
|
|
|
|
|
|
|
|
o << "#ifdef ENABLE_INSTR_PREDICATE_VERIFIER\n"
|
|
|
|
<< "#undef ENABLE_INSTR_PREDICATE_VERIFIER\n"
|
|
|
|
<< "#include <sstream>\n\n";
|
|
|
|
|
|
|
|
// Emit the subtarget feature enumeration.
|
|
|
|
SubtargetFeatureInfo::emitSubtargetFeatureFlagEnumeration(SubtargetFeatures,
|
|
|
|
o);
|
|
|
|
|
|
|
|
// Emit the name table for error messages.
|
|
|
|
o << "#ifndef NDEBUG\n";
|
|
|
|
SubtargetFeatureInfo::emitNameTable(SubtargetFeatures, o);
|
|
|
|
o << "#endif // NDEBUG\n";
|
|
|
|
|
|
|
|
// Emit the available features compute function.
|
[globalisel][tablegen] Import SelectionDAG's rule predicates and support the equivalent in GIRule.
Summary:
The SelectionDAG importer now imports rules with Predicate's attached via
Requires, PredicateControl, etc. These predicates are implemented as
bitset's to allow multiple predicates to be tested together. However,
unlike the MC layer subtarget features, each target only pays for it's own
predicates (e.g. AArch64 doesn't have 192 feature bits just because X86
needs a lot).
Both AArch64 and X86 derive at least one predicate from the MachineFunction
or Function so they must re-initialize AvailableFeatures before each
function. They also declare locals in <Target>InstructionSelector so that
computeAvailableFeatures() can use the code from SelectionDAG without
modification.
Reviewers: rovka, qcolombet, aditya_nandakumar, t.p.northover, ab
Reviewed By: rovka
Subscribers: aemerson, rengolin, dberris, kristof.beyls, llvm-commits, igorb
Differential Revision: https://reviews.llvm.org/D31418
llvm-svn: 300993
2017-04-21 23:59:56 +08:00
|
|
|
SubtargetFeatureInfo::emitComputeAssemblerAvailableFeatures(
|
Check that emitted instructions meet their predicates on all targets except ARM, Mips, and X86.
Summary:
* ARM is omitted from this patch because this check appears to expose bugs in this target.
* Mips is omitted from this patch because this check either detects bugs or deliberate
emission of instructions that don't satisfy their predicates. One deliberate
use is the SYNC instruction where the version with an operand is correctly
defined as requiring MIPS32 while the version without an operand is defined
as an alias of 'SYNC 0' and requires MIPS2.
* X86 is omitted from this patch because it doesn't use the tablegen-erated
MCCodeEmitter infrastructure.
Patches for ARM and Mips will follow.
Depends on D25617
Reviewers: tstellarAMD, jmolloy
Subscribers: wdng, jmolloy, aemerson, rengolin, arsenm, jyknight, nemanjai, nhaehnle, tstellarAMD, llvm-commits
Differential Revision: https://reviews.llvm.org/D25618
llvm-svn: 287439
2016-11-19 21:05:44 +08:00
|
|
|
Target.getName(), "MCCodeEmitter", "computeAvailableFeatures",
|
|
|
|
SubtargetFeatures, o);
|
|
|
|
|
|
|
|
// Emit the predicate verifier.
|
|
|
|
o << "void " << Target.getName()
|
|
|
|
<< "MCCodeEmitter::verifyInstructionPredicates(\n"
|
|
|
|
<< " const MCInst &Inst, uint64_t AvailableFeatures) const {\n"
|
|
|
|
<< "#ifndef NDEBUG\n"
|
|
|
|
<< " static uint64_t RequiredFeatures[] = {\n";
|
|
|
|
unsigned InstIdx = 0;
|
|
|
|
for (const CodeGenInstruction *Inst : Target.getInstructionsByEnumValue()) {
|
|
|
|
o << " ";
|
|
|
|
for (Record *Predicate : Inst->TheDef->getValueAsListOfDefs("Predicates")) {
|
|
|
|
const auto &I = SubtargetFeatures.find(Predicate);
|
|
|
|
if (I != SubtargetFeatures.end())
|
|
|
|
o << I->second.getEnumName() << " | ";
|
|
|
|
}
|
|
|
|
o << "0, // " << Inst->TheDef->getName() << " = " << InstIdx << "\n";
|
|
|
|
InstIdx++;
|
|
|
|
}
|
|
|
|
o << " };\n\n";
|
|
|
|
o << " assert(Inst.getOpcode() < " << InstIdx << ");\n";
|
|
|
|
o << " uint64_t MissingFeatures =\n"
|
|
|
|
<< " (AvailableFeatures & RequiredFeatures[Inst.getOpcode()]) ^\n"
|
|
|
|
<< " RequiredFeatures[Inst.getOpcode()];\n"
|
|
|
|
<< " if (MissingFeatures) {\n"
|
|
|
|
<< " std::ostringstream Msg;\n"
|
2016-11-19 22:47:41 +08:00
|
|
|
<< " Msg << \"Attempting to emit \" << "
|
|
|
|
"MCII.getName(Inst.getOpcode()).str()\n"
|
Check that emitted instructions meet their predicates on all targets except ARM, Mips, and X86.
Summary:
* ARM is omitted from this patch because this check appears to expose bugs in this target.
* Mips is omitted from this patch because this check either detects bugs or deliberate
emission of instructions that don't satisfy their predicates. One deliberate
use is the SYNC instruction where the version with an operand is correctly
defined as requiring MIPS32 while the version without an operand is defined
as an alias of 'SYNC 0' and requires MIPS2.
* X86 is omitted from this patch because it doesn't use the tablegen-erated
MCCodeEmitter infrastructure.
Patches for ARM and Mips will follow.
Depends on D25617
Reviewers: tstellarAMD, jmolloy
Subscribers: wdng, jmolloy, aemerson, rengolin, arsenm, jyknight, nemanjai, nhaehnle, tstellarAMD, llvm-commits
Differential Revision: https://reviews.llvm.org/D25618
llvm-svn: 287439
2016-11-19 21:05:44 +08:00
|
|
|
<< " << \" instruction but the \";\n"
|
|
|
|
<< " for (unsigned i = 0; i < 8 * sizeof(MissingFeatures); ++i)\n"
|
|
|
|
<< " if (MissingFeatures & (1ULL << i))\n"
|
|
|
|
<< " Msg << SubtargetFeatureNames[i] << \" \";\n"
|
|
|
|
<< " Msg << \"predicate(s) are not met\";\n"
|
|
|
|
<< " report_fatal_error(Msg.str());\n"
|
|
|
|
<< " }\n"
|
2016-11-19 22:47:41 +08:00
|
|
|
<< "#else\n"
|
|
|
|
<< "// Silence unused variable warning on targets that don't use MCII for "
|
|
|
|
"other purposes (e.g. BPF).\n"
|
|
|
|
<< "(void)MCII;\n"
|
Check that emitted instructions meet their predicates on all targets except ARM, Mips, and X86.
Summary:
* ARM is omitted from this patch because this check appears to expose bugs in this target.
* Mips is omitted from this patch because this check either detects bugs or deliberate
emission of instructions that don't satisfy their predicates. One deliberate
use is the SYNC instruction where the version with an operand is correctly
defined as requiring MIPS32 while the version without an operand is defined
as an alias of 'SYNC 0' and requires MIPS2.
* X86 is omitted from this patch because it doesn't use the tablegen-erated
MCCodeEmitter infrastructure.
Patches for ARM and Mips will follow.
Depends on D25617
Reviewers: tstellarAMD, jmolloy
Subscribers: wdng, jmolloy, aemerson, rengolin, arsenm, jyknight, nemanjai, nhaehnle, tstellarAMD, llvm-commits
Differential Revision: https://reviews.llvm.org/D25618
llvm-svn: 287439
2016-11-19 21:05:44 +08:00
|
|
|
<< "#endif // NDEBUG\n";
|
|
|
|
o << "}\n";
|
|
|
|
o << "#endif\n";
|
2003-10-06 03:27:59 +08:00
|
|
|
}
|
2012-06-11 23:37:55 +08:00
|
|
|
|
2016-12-01 01:48:10 +08:00
|
|
|
} // end anonymous namespace
|
2012-06-11 23:37:55 +08:00
|
|
|
|
|
|
|
namespace llvm {
|
|
|
|
|
|
|
|
void EmitCodeEmitter(RecordKeeper &RK, raw_ostream &OS) {
|
|
|
|
emitSourceFileHeader("Machine Code Emitter", OS);
|
|
|
|
CodeEmitterGen(RK).run(OS);
|
|
|
|
}
|
|
|
|
|
2016-12-01 01:48:10 +08:00
|
|
|
} // end namespace llvm
|