forked from OSchip/llvm-project
[modules] Preserve DeclID order when merging lookup tables to give a more
predictable diagnostic experience. The hash-of-DeclID order we were using before gave different results on Win32 due to a different predefined declaration of __builtin_va_list. llvm-svn: 246521
This commit is contained in:
parent
a8dde6c83f
commit
7fa450c6ae
|
@ -959,7 +959,7 @@ ASTDeclContextNameLookupTrait::ReadKey(const unsigned char *d, unsigned) {
|
|||
void ASTDeclContextNameLookupTrait::ReadDataInto(internal_key_type,
|
||||
const unsigned char *d,
|
||||
unsigned DataLen,
|
||||
data_type &Val) {
|
||||
data_type_builder &Val) {
|
||||
using namespace llvm::support;
|
||||
for (unsigned NumDecls = DataLen / 4; NumDecls; --NumDecls) {
|
||||
uint32_t LocalID = endian::readNext<uint32_t, little, unaligned>(d);
|
||||
|
|
|
@ -47,8 +47,31 @@ public:
|
|||
static const int MaxTables = 4;
|
||||
|
||||
/// The lookup result is a list of global declaration IDs.
|
||||
// FIXME: LLVM doesn't really have a good data structure for this.
|
||||
typedef llvm::DenseSet<DeclID> data_type;
|
||||
typedef llvm::SmallVector<DeclID, 4> data_type;
|
||||
struct data_type_builder {
|
||||
data_type &Data;
|
||||
llvm::DenseSet<DeclID> Found;
|
||||
|
||||
data_type_builder(data_type &D) : Data(D) {}
|
||||
void insert(DeclID ID) {
|
||||
// Just use a linear scan unless we have more than a few IDs.
|
||||
if (Found.empty() && !Data.empty()) {
|
||||
if (Data.size() <= 4) {
|
||||
for (auto I : Found)
|
||||
if (I == ID)
|
||||
return;
|
||||
Data.push_back(ID);
|
||||
return;
|
||||
}
|
||||
|
||||
// Switch to tracking found IDs in the set.
|
||||
Found.insert(Data.begin(), Data.end());
|
||||
}
|
||||
|
||||
if (Found.insert(ID).second)
|
||||
Data.push_back(ID);
|
||||
}
|
||||
};
|
||||
typedef unsigned hash_value_type;
|
||||
typedef unsigned offset_type;
|
||||
typedef ModuleFile *file_type;
|
||||
|
@ -76,10 +99,12 @@ public:
|
|||
internal_key_type ReadKey(const unsigned char *d, unsigned);
|
||||
|
||||
void ReadDataInto(internal_key_type, const unsigned char *d,
|
||||
unsigned DataLen, data_type &Val);
|
||||
unsigned DataLen, data_type_builder &Val);
|
||||
|
||||
static void MergeDataInto(const data_type &From, data_type &To) {
|
||||
To.insert(From.begin(), From.end());
|
||||
static void MergeDataInto(const data_type &From, data_type_builder &To) {
|
||||
To.Data.reserve(To.Data.size() + From.size());
|
||||
for (DeclID ID : From)
|
||||
To.insert(ID);
|
||||
}
|
||||
|
||||
file_type ReadFileRef(const unsigned char *&d);
|
||||
|
|
|
@ -38,6 +38,7 @@ public:
|
|||
typedef typename Info::external_key_type external_key_type;
|
||||
typedef typename Info::internal_key_type internal_key_type;
|
||||
typedef typename Info::data_type data_type;
|
||||
typedef typename Info::data_type_builder data_type_builder;
|
||||
typedef unsigned hash_value_type;
|
||||
|
||||
private:
|
||||
|
@ -135,8 +136,9 @@ private:
|
|||
// FIXME: Don't rely on the OnDiskHashTable format here.
|
||||
auto L = InfoObj.ReadKeyDataLength(LocalPtr);
|
||||
const internal_key_type &Key = InfoObj.ReadKey(LocalPtr, L.first);
|
||||
data_type_builder ValueBuilder(Merged->Data[Key]);
|
||||
InfoObj.ReadDataInto(Key, LocalPtr + L.first, L.second,
|
||||
Merged->Data[Key]);
|
||||
ValueBuilder);
|
||||
}
|
||||
|
||||
Merged->Files.push_back(ODT->File);
|
||||
|
@ -218,12 +220,14 @@ public:
|
|||
Result = It->second;
|
||||
}
|
||||
|
||||
data_type_builder ResultBuilder(Result);
|
||||
|
||||
for (auto *ODT : tables()) {
|
||||
auto &HT = ODT->Table;
|
||||
auto It = HT.find_hashed(Key, KeyHash);
|
||||
if (It != HT.end())
|
||||
HT.getInfoObj().ReadDataInto(Key, It.getDataPtr(), It.getDataLen(),
|
||||
Result);
|
||||
ResultBuilder);
|
||||
}
|
||||
|
||||
return Result;
|
||||
|
@ -233,13 +237,14 @@ public:
|
|||
/// sense if merging values across keys is meaningful.
|
||||
data_type findAll() {
|
||||
data_type Result;
|
||||
data_type_builder ResultBuilder(Result);
|
||||
|
||||
if (!PendingOverrides.empty())
|
||||
removeOverriddenTables();
|
||||
|
||||
if (MergedTable *M = getMergedTable()) {
|
||||
for (auto &KV : M->Data)
|
||||
Info::MergeDataInto(KV.second, Result);
|
||||
Info::MergeDataInto(KV.second, ResultBuilder);
|
||||
}
|
||||
|
||||
for (auto *ODT : tables()) {
|
||||
|
@ -251,7 +256,7 @@ public:
|
|||
// FIXME: Don't rely on the OnDiskHashTable format here.
|
||||
auto L = InfoObj.ReadKeyDataLength(LocalPtr);
|
||||
const internal_key_type &Key = InfoObj.ReadKey(LocalPtr, L.first);
|
||||
InfoObj.ReadDataInto(Key, LocalPtr + L.first, L.second, Result);
|
||||
InfoObj.ReadDataInto(Key, LocalPtr + L.first, L.second, ResultBuilder);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -28,15 +28,15 @@ void g() {
|
|||
f<double>(1.0);
|
||||
f<int>();
|
||||
f(); // expected-error {{no matching function}}
|
||||
// expected-note@Inputs/cxx-templates-b.h:3 {{couldn't infer template argument}}
|
||||
// expected-note-re@Inputs/cxx-templates-a.h:4 {{requires {{single|1}} argument}}
|
||||
// expected-note@Inputs/cxx-templates-a.h:3 {{couldn't infer template argument}}
|
||||
// expected-note@Inputs/cxx-templates-a.h:4 {{requires 1 argument}}
|
||||
|
||||
N::f(0);
|
||||
N::f<double>(1.0);
|
||||
N::f<int>();
|
||||
N::f(); // expected-error {{no matching function}}
|
||||
// expected-note@Inputs/cxx-templates-b.h:6 {{couldn't infer template argument}}
|
||||
// expected-note-re@Inputs/cxx-templates-a.h:7 {{requires {{single|1}} argument}}
|
||||
// expected-note@Inputs/cxx-templates-b.h:7 {{requires single argument}}
|
||||
|
||||
template_param_kinds_1<0>(); // ok, from cxx-templates-a.h
|
||||
template_param_kinds_1<int>(); // ok, from cxx-templates-b.h
|
||||
|
|
|
@ -33,7 +33,7 @@ template int UseAll<Y>();
|
|||
|
||||
// Which of these two sets of diagnostics is chosen is not important. It's OK
|
||||
// if this varies with ORDER, but it must be consistent across runs.
|
||||
#if 1
|
||||
#if ORDER == 1
|
||||
// Here, we're instantiating the definition from 'A' and merging the definition
|
||||
// from 'B' into it.
|
||||
|
||||
|
|
Loading…
Reference in New Issue