forked from OSchip/llvm-project
456 lines
17 KiB
C++
456 lines
17 KiB
C++
//===--- OptimizedStructLayout.cpp - Optimal data layout algorithm ----------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
//
|
|
// This file implements the performOptimizedStructLayout interface.
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "llvm/Support/OptimizedStructLayout.h"
|
|
|
|
using namespace llvm;
|
|
|
|
using Field = OptimizedStructLayoutField;
|
|
|
|
#ifndef NDEBUG
|
|
static void checkValidLayout(ArrayRef<Field> Fields, uint64_t Size,
|
|
Align MaxAlign) {
|
|
uint64_t LastEnd = 0;
|
|
Align ComputedMaxAlign;
|
|
for (auto &Field : Fields) {
|
|
assert(Field.hasFixedOffset() &&
|
|
"didn't assign a fixed offset to field");
|
|
assert(isAligned(Field.Alignment, Field.Offset) &&
|
|
"didn't assign a correctly-aligned offset to field");
|
|
assert(Field.Offset >= LastEnd &&
|
|
"didn't assign offsets in ascending order");
|
|
LastEnd = Field.getEndOffset();
|
|
assert(Field.Alignment <= MaxAlign &&
|
|
"didn't compute MaxAlign correctly");
|
|
ComputedMaxAlign = std::max(Field.Alignment, MaxAlign);
|
|
}
|
|
assert(LastEnd == Size && "didn't compute LastEnd correctly");
|
|
assert(ComputedMaxAlign == MaxAlign && "didn't compute MaxAlign correctly");
|
|
}
|
|
#endif
|
|
|
|
std::pair<uint64_t, Align>
|
|
llvm::performOptimizedStructLayout(MutableArrayRef<Field> Fields) {
|
|
#ifndef NDEBUG
|
|
// Do some simple precondition checks.
|
|
{
|
|
bool InFixedPrefix = true;
|
|
size_t LastEnd = 0;
|
|
for (auto &Field : Fields) {
|
|
assert(Field.Size > 0 && "field of zero size");
|
|
if (Field.hasFixedOffset()) {
|
|
assert(InFixedPrefix &&
|
|
"fixed-offset fields are not a strict prefix of array");
|
|
assert(LastEnd <= Field.Offset &&
|
|
"fixed-offset fields overlap or are not in order");
|
|
LastEnd = Field.getEndOffset();
|
|
assert(LastEnd > Field.Offset &&
|
|
"overflow in fixed-offset end offset");
|
|
} else {
|
|
InFixedPrefix = false;
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
|
|
// Do an initial pass over the fields.
|
|
Align MaxAlign;
|
|
|
|
// Find the first flexible-offset field, tracking MaxAlign.
|
|
auto FirstFlexible = Fields.begin(), E = Fields.end();
|
|
while (FirstFlexible != E && FirstFlexible->hasFixedOffset()) {
|
|
MaxAlign = std::max(MaxAlign, FirstFlexible->Alignment);
|
|
++FirstFlexible;
|
|
}
|
|
|
|
// If there are no flexible fields, we're done.
|
|
if (FirstFlexible == E) {
|
|
uint64_t Size = 0;
|
|
if (!Fields.empty())
|
|
Size = Fields.back().getEndOffset();
|
|
|
|
#ifndef NDEBUG
|
|
checkValidLayout(Fields, Size, MaxAlign);
|
|
#endif
|
|
return std::make_pair(Size, MaxAlign);
|
|
}
|
|
|
|
// Walk over the flexible-offset fields, tracking MaxAlign and
|
|
// assigning them a unique number in order of their appearance.
|
|
// We'll use this unique number in the comparison below so that
|
|
// we can use array_pod_sort, which isn't stable. We won't use it
|
|
// past that point.
|
|
{
|
|
uintptr_t UniqueNumber = 0;
|
|
for (auto I = FirstFlexible; I != E; ++I) {
|
|
I->Scratch = reinterpret_cast<void*>(UniqueNumber++);
|
|
MaxAlign = std::max(MaxAlign, I->Alignment);
|
|
}
|
|
}
|
|
|
|
// Sort the flexible elements in order of decreasing alignment,
|
|
// then decreasing size, and then the original order as recorded
|
|
// in Scratch. The decreasing-size aspect of this is only really
|
|
// important if we get into the gap-filling stage below, but it
|
|
// doesn't hurt here.
|
|
array_pod_sort(FirstFlexible, E,
|
|
[](const Field *lhs, const Field *rhs) -> int {
|
|
// Decreasing alignment.
|
|
if (lhs->Alignment != rhs->Alignment)
|
|
return (lhs->Alignment < rhs->Alignment ? 1 : -1);
|
|
|
|
// Decreasing size.
|
|
if (lhs->Size != rhs->Size)
|
|
return (lhs->Size < rhs->Size ? 1 : -1);
|
|
|
|
// Original order.
|
|
auto lhsNumber = reinterpret_cast<uintptr_t>(lhs->Scratch);
|
|
auto rhsNumber = reinterpret_cast<uintptr_t>(rhs->Scratch);
|
|
if (lhsNumber != rhsNumber)
|
|
return (lhsNumber < rhsNumber ? -1 : 1);
|
|
|
|
return 0;
|
|
});
|
|
|
|
// Do a quick check for whether that sort alone has given us a perfect
|
|
// layout with no interior padding. This is very common: if the
|
|
// fixed-layout fields have no interior padding, and they end at a
|
|
// sufficiently-aligned offset for all the flexible-layout fields,
|
|
// and the flexible-layout fields all have sizes that are multiples
|
|
// of their alignment, then this will reliably trigger.
|
|
{
|
|
bool HasPadding = false;
|
|
uint64_t LastEnd = 0;
|
|
|
|
// Walk the fixed-offset fields.
|
|
for (auto I = Fields.begin(); I != FirstFlexible; ++I) {
|
|
assert(I->hasFixedOffset());
|
|
if (LastEnd != I->Offset) {
|
|
HasPadding = true;
|
|
break;
|
|
}
|
|
LastEnd = I->getEndOffset();
|
|
}
|
|
|
|
// Walk the flexible-offset fields, optimistically assigning fixed
|
|
// offsets. Note that we maintain a strict division between the
|
|
// fixed-offset and flexible-offset fields, so if we end up
|
|
// discovering padding later in this loop, we can just abandon this
|
|
// work and we'll ignore the offsets we already assigned.
|
|
if (!HasPadding) {
|
|
for (auto I = FirstFlexible; I != E; ++I) {
|
|
auto Offset = alignTo(LastEnd, I->Alignment);
|
|
if (LastEnd != Offset) {
|
|
HasPadding = true;
|
|
break;
|
|
}
|
|
I->Offset = Offset;
|
|
LastEnd = I->getEndOffset();
|
|
}
|
|
}
|
|
|
|
// If we already have a perfect layout, we're done.
|
|
if (!HasPadding) {
|
|
#ifndef NDEBUG
|
|
checkValidLayout(Fields, LastEnd, MaxAlign);
|
|
#endif
|
|
return std::make_pair(LastEnd, MaxAlign);
|
|
}
|
|
}
|
|
|
|
// The algorithm sketch at this point is as follows.
|
|
//
|
|
// Consider the padding gaps between fixed-offset fields in ascending
|
|
// order. Let LastEnd be the offset of the first byte following the
|
|
// field before the gap, or 0 if the gap is at the beginning of the
|
|
// structure. Find the "best" flexible-offset field according to the
|
|
// criteria below. If no such field exists, proceed to the next gap.
|
|
// Otherwise, add the field at the first properly-aligned offset for
|
|
// that field that is >= LastEnd, then update LastEnd and repeat in
|
|
// order to fill any remaining gap following that field.
|
|
//
|
|
// Next, let LastEnd to be the offset of the first byte following the
|
|
// last fixed-offset field, or 0 if there are no fixed-offset fields.
|
|
// While there are flexible-offset fields remaining, find the "best"
|
|
// flexible-offset field according to the criteria below, add it at
|
|
// the first properly-aligned offset for that field that is >= LastEnd,
|
|
// and update LastEnd to the first byte following the field.
|
|
//
|
|
// The "best" field is chosen by the following criteria, considered
|
|
// strictly in order:
|
|
//
|
|
// - When filling a gap betweeen fields, the field must fit.
|
|
// - A field is preferred if it requires less padding following LastEnd.
|
|
// - A field is preferred if it is more aligned.
|
|
// - A field is preferred if it is larger.
|
|
// - A field is preferred if it appeared earlier in the initial order.
|
|
//
|
|
// Minimizing leading padding is a greedy attempt to avoid padding
|
|
// entirely. Preferring more-aligned fields is an attempt to eliminate
|
|
// stricter constraints earlier, with the idea that weaker alignment
|
|
// constraints may be resolvable with less padding elsewhere. These
|
|
// These two rules are sufficient to ensure that we get the optimal
|
|
// layout in the "C-style" case. Preferring larger fields tends to take
|
|
// better advantage of large gaps and may be more likely to have a size
|
|
// that's a multiple of a useful alignment. Preferring the initial
|
|
// order may help somewhat with locality but is mostly just a way of
|
|
// ensuring deterministic output.
|
|
//
|
|
// Note that this algorithm does not guarantee a minimal layout. Picking
|
|
// a larger object greedily may leave a gap that cannot be filled as
|
|
// efficiently. Unfortunately, solving this perfectly is an NP-complete
|
|
// problem (by reduction from bin-packing: let B_i be the bin sizes and
|
|
// O_j be the object sizes; add fixed-offset fields such that the gaps
|
|
// between them have size B_i, and add flexible-offset fields with
|
|
// alignment 1 and size O_j; if the layout size is equal to the end of
|
|
// the last fixed-layout field, the objects fit in the bins; note that
|
|
// this doesn't even require the complexity of alignment).
|
|
|
|
// The implementation below is essentially just an optimized version of
|
|
// scanning the list of remaining fields looking for the best, which
|
|
// would be O(n^2). In the worst case, it doesn't improve on that.
|
|
// However, in practice it'll just scan the array of alignment bins
|
|
// and consider the first few elements from one or two bins. The
|
|
// number of bins is bounded by a small constant: alignments are powers
|
|
// of two that are vanishingly unlikely to be over 64 and fairly unlikely
|
|
// to be over 8. And multiple elements only need to be considered when
|
|
// filling a gap between fixed-offset fields, which doesn't happen very
|
|
// often. We could use a data structure within bins that optimizes for
|
|
// finding the best-sized match, but it would require allocating memory
|
|
// and copying data, so it's unlikely to be worthwhile.
|
|
|
|
|
|
// Start by organizing the flexible-offset fields into bins according to
|
|
// their alignment. We expect a small enough number of bins that we
|
|
// don't care about the asymptotic costs of walking this.
|
|
struct AlignmentQueue {
|
|
/// The minimum size of anything currently in this queue.
|
|
uint64_t MinSize;
|
|
|
|
/// The head of the queue. A singly-linked list. The order here should
|
|
/// be consistent with the earlier sort, i.e. the elements should be
|
|
/// monotonically descending in size and otherwise in the original order.
|
|
///
|
|
/// We remove the queue from the array as soon as this is empty.
|
|
OptimizedStructLayoutField *Head;
|
|
|
|
/// The alignment requirement of the queue.
|
|
Align Alignment;
|
|
|
|
static Field *getNext(Field *Cur) {
|
|
return static_cast<Field *>(Cur->Scratch);
|
|
}
|
|
};
|
|
SmallVector<AlignmentQueue, 8> FlexibleFieldsByAlignment;
|
|
for (auto I = FirstFlexible; I != E; ) {
|
|
auto Head = I;
|
|
auto Alignment = I->Alignment;
|
|
|
|
uint64_t MinSize = I->Size;
|
|
auto LastInQueue = I;
|
|
for (++I; I != E && I->Alignment == Alignment; ++I) {
|
|
LastInQueue->Scratch = I;
|
|
LastInQueue = I;
|
|
MinSize = std::min(MinSize, I->Size);
|
|
}
|
|
LastInQueue->Scratch = nullptr;
|
|
|
|
FlexibleFieldsByAlignment.push_back({MinSize, Head, Alignment});
|
|
}
|
|
|
|
#ifndef NDEBUG
|
|
// Verify that we set the queues up correctly.
|
|
auto checkQueues = [&]{
|
|
bool FirstQueue = true;
|
|
Align LastQueueAlignment;
|
|
for (auto &Queue : FlexibleFieldsByAlignment) {
|
|
assert((FirstQueue || Queue.Alignment < LastQueueAlignment) &&
|
|
"bins not in order of descending alignment");
|
|
LastQueueAlignment = Queue.Alignment;
|
|
FirstQueue = false;
|
|
|
|
assert(Queue.Head && "queue was empty");
|
|
uint64_t LastSize = ~(uint64_t)0;
|
|
for (auto I = Queue.Head; I; I = Queue.getNext(I)) {
|
|
assert(I->Alignment == Queue.Alignment && "bad field in queue");
|
|
assert(I->Size <= LastSize && "queue not in descending size order");
|
|
LastSize = I->Size;
|
|
}
|
|
}
|
|
};
|
|
checkQueues();
|
|
#endif
|
|
|
|
/// Helper function to remove a field from a queue.
|
|
auto spliceFromQueue = [&](AlignmentQueue *Queue, Field *Last, Field *Cur) {
|
|
assert(Last ? Queue->getNext(Last) == Cur : Queue->Head == Cur);
|
|
|
|
// If we're removing Cur from a non-initial position, splice it out
|
|
// of the linked list.
|
|
if (Last) {
|
|
Last->Scratch = Cur->Scratch;
|
|
|
|
// If Cur was the last field in the list, we need to update MinSize.
|
|
// We can just use the last field's size because the list is in
|
|
// descending order of size.
|
|
if (!Cur->Scratch)
|
|
Queue->MinSize = Last->Size;
|
|
|
|
// Otherwise, replace the head.
|
|
} else {
|
|
if (auto NewHead = Queue->getNext(Cur))
|
|
Queue->Head = NewHead;
|
|
|
|
// If we just emptied the queue, destroy its bin.
|
|
else
|
|
FlexibleFieldsByAlignment.erase(Queue);
|
|
}
|
|
};
|
|
|
|
// Do layout into a local array. Doing this in-place on Fields is
|
|
// not really feasible.
|
|
SmallVector<Field, 16> Layout;
|
|
Layout.reserve(Fields.size());
|
|
|
|
// The offset that we're currently looking to insert at (or after).
|
|
uint64_t LastEnd = 0;
|
|
|
|
// Helper function to splice Cur out of the given queue and add it
|
|
// to the layout at the given offset.
|
|
auto addToLayout = [&](AlignmentQueue *Queue, Field *Last, Field *Cur,
|
|
uint64_t Offset) -> bool {
|
|
assert(Offset == alignTo(LastEnd, Cur->Alignment));
|
|
|
|
// Splice out. This potentially invalidates Queue.
|
|
spliceFromQueue(Queue, Last, Cur);
|
|
|
|
// Add Cur to the layout.
|
|
Layout.push_back(*Cur);
|
|
Layout.back().Offset = Offset;
|
|
LastEnd = Layout.back().getEndOffset();
|
|
|
|
// Always return true so that we can be tail-called.
|
|
return true;
|
|
};
|
|
|
|
// Helper function to try to find a field in the given queue that'll
|
|
// fit starting at StartOffset but before EndOffset (if present).
|
|
// Note that this never fails if EndOffset is not provided.
|
|
auto tryAddFillerFromQueue = [&](AlignmentQueue *Queue,
|
|
uint64_t StartOffset,
|
|
Optional<uint64_t> EndOffset) -> bool {
|
|
assert(Queue->Head);
|
|
assert(StartOffset == alignTo(LastEnd, Queue->Alignment));
|
|
assert(!EndOffset || StartOffset < *EndOffset);
|
|
|
|
// Figure out the maximum size that a field can be, and ignore this
|
|
// queue if there's nothing in it that small.
|
|
auto MaxViableSize =
|
|
(EndOffset ? *EndOffset - StartOffset : ~(uint64_t)0);
|
|
if (Queue->MinSize > MaxViableSize) return false;
|
|
|
|
// Find the matching field. Note that this should always find
|
|
// something because of the MinSize check above.
|
|
for (Field *Cur = Queue->Head, *Last = nullptr; true;
|
|
Last = Cur, Cur = Queue->getNext(Cur)) {
|
|
assert(Cur && "didn't find a match in queue despite its MinSize");
|
|
if (Cur->Size <= MaxViableSize)
|
|
return addToLayout(Queue, Last, Cur, StartOffset);
|
|
}
|
|
|
|
llvm_unreachable("didn't find a match in queue despite its MinSize");
|
|
};
|
|
|
|
// Helper function to find the "best" flexible-offset field according
|
|
// to the criteria described above.
|
|
auto tryAddBestField = [&](Optional<uint64_t> BeforeOffset) -> bool {
|
|
assert(!BeforeOffset || LastEnd < *BeforeOffset);
|
|
auto QueueB = FlexibleFieldsByAlignment.begin();
|
|
auto QueueE = FlexibleFieldsByAlignment.end();
|
|
|
|
// Start by looking for the most-aligned queue that doesn't need any
|
|
// leading padding after LastEnd.
|
|
auto FirstQueueToSearch = QueueB;
|
|
for (; FirstQueueToSearch != QueueE; ++FirstQueueToSearch) {
|
|
if (isAligned(FirstQueueToSearch->Alignment, LastEnd))
|
|
break;
|
|
}
|
|
|
|
uint64_t Offset = LastEnd;
|
|
while (true) {
|
|
// Invariant: all of the queues in [FirstQueueToSearch, QueueE)
|
|
// require the same initial padding offset.
|
|
|
|
// Search those queues in descending order of alignment for a
|
|
// satisfactory field.
|
|
for (auto Queue = FirstQueueToSearch; Queue != QueueE; ++Queue) {
|
|
if (tryAddFillerFromQueue(Queue, Offset, BeforeOffset))
|
|
return true;
|
|
}
|
|
|
|
// Okay, we don't need to scan those again.
|
|
QueueE = FirstQueueToSearch;
|
|
|
|
// If we started from the first queue, we're done.
|
|
if (FirstQueueToSearch == QueueB)
|
|
return false;
|
|
|
|
// Otherwise, scan backwards to find the most-aligned queue that
|
|
// still has minimal leading padding after LastEnd. If that
|
|
// minimal padding is already at or past the end point, we're done.
|
|
--FirstQueueToSearch;
|
|
Offset = alignTo(LastEnd, FirstQueueToSearch->Alignment);
|
|
if (BeforeOffset && Offset >= *BeforeOffset)
|
|
return false;
|
|
while (FirstQueueToSearch != QueueB &&
|
|
Offset == alignTo(LastEnd, FirstQueueToSearch[-1].Alignment))
|
|
--FirstQueueToSearch;
|
|
}
|
|
};
|
|
|
|
// Phase 1: fill the gaps between fixed-offset fields with the best
|
|
// flexible-offset field that fits.
|
|
for (auto I = Fields.begin(); I != FirstFlexible; ++I) {
|
|
assert(LastEnd <= I->Offset);
|
|
while (LastEnd != I->Offset) {
|
|
if (!tryAddBestField(I->Offset))
|
|
break;
|
|
}
|
|
Layout.push_back(*I);
|
|
LastEnd = I->getEndOffset();
|
|
}
|
|
|
|
#ifndef NDEBUG
|
|
checkQueues();
|
|
#endif
|
|
|
|
// Phase 2: repeatedly add the best flexible-offset field until
|
|
// they're all gone.
|
|
while (!FlexibleFieldsByAlignment.empty()) {
|
|
bool Success = tryAddBestField(None);
|
|
assert(Success && "didn't find a field with no fixed limit?");
|
|
(void) Success;
|
|
}
|
|
|
|
// Copy the layout back into place.
|
|
assert(Layout.size() == Fields.size());
|
|
memcpy(Fields.data(), Layout.data(),
|
|
Fields.size() * sizeof(OptimizedStructLayoutField));
|
|
|
|
#ifndef NDEBUG
|
|
// Make a final check that the layout is valid.
|
|
checkValidLayout(Fields, LastEnd, MaxAlign);
|
|
#endif
|
|
|
|
return std::make_pair(LastEnd, MaxAlign);
|
|
}
|