1103 lines
40 KiB
C++
1103 lines
40 KiB
C++
/*
|
|
* DiskQueue.actor.cpp
|
|
*
|
|
* This source file is part of the FoundationDB open source project
|
|
*
|
|
* Copyright 2013-2018 Apple Inc. and the FoundationDB project authors
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include "flow/actorcompiler.h"
|
|
#include "IDiskQueue.h"
|
|
#include "fdbrpc/IAsyncFile.h"
|
|
#include "Knobs.h"
|
|
#include "fdbrpc/simulator.h"
|
|
|
|
typedef bool(*compare_pages)(void*,void*);
|
|
typedef int64_t loc_t;
|
|
|
|
struct StringBuffer {
|
|
Standalone<StringRef> str;
|
|
int reserved;
|
|
UID id;
|
|
|
|
StringBuffer(UID fromFileID) : reserved(0), id( fromFileID ) {}
|
|
|
|
int size() const { return str.size(); }
|
|
StringRef& ref() { return str; }
|
|
void clear() {
|
|
str = Standalone<StringRef>();
|
|
reserved = 0;
|
|
}
|
|
void clearReserve(int size) {
|
|
str = Standalone<StringRef>();
|
|
reserved = size;
|
|
ref() = StringRef( new (str.arena()) uint8_t[size], 0 );
|
|
}
|
|
void append( StringRef x ) {
|
|
memcpy( append(x.size()), x.begin(), x.size() );
|
|
}
|
|
void* append(int bytes) {
|
|
ASSERT( str.size() + bytes <= reserved );
|
|
void* p = const_cast<uint8_t*>(str.end());
|
|
ref() = StringRef( str.begin(), str.size()+bytes );
|
|
return p;
|
|
}
|
|
StringRef pop_front(int bytes) {
|
|
ASSERT( bytes <= str.size() );
|
|
StringRef result = str.substr(0, bytes);
|
|
ref() = str.substr(bytes);
|
|
return result;
|
|
}
|
|
void alignReserve(int alignment, int size) {
|
|
ASSERT( alignment && (alignment & (alignment-1)) == 0 ); // alignment is a power of two
|
|
|
|
if (size >= reserved) {
|
|
// SOMEDAY: Use a new arena and discard the old one after copying?
|
|
reserved = std::max( size, reserved*2 );
|
|
if( reserved > 1e9 ) {
|
|
printf("WOAH! Huge allocation\n");
|
|
TraceEvent(SevError, "StringBufferHugeAllocation", id).detail("Alignment", alignment).detail("Reserved", reserved).backtrace();
|
|
}
|
|
uint8_t* b = new (str.arena()) uint8_t[reserved+alignment-1];
|
|
uint8_t* e = b + (reserved+alignment-1);
|
|
|
|
uint8_t* p = (uint8_t*)(int64_t(b+alignment-1) & ~(alignment-1)); // first multiple of alignment greater than or equal to b
|
|
ASSERT( p>=b && p+reserved<=e && int64_t(p)%alignment == 0 );
|
|
|
|
memcpy(p, str.begin(), str.size());
|
|
ref() = StringRef( p, str.size() );
|
|
}
|
|
}
|
|
};
|
|
|
|
struct SyncQueue : ReferenceCounted<SyncQueue> {
|
|
SyncQueue( int outstandingLimit, Reference<IAsyncFile> file )
|
|
: outstandingLimit(outstandingLimit), file(file)
|
|
{
|
|
for(int i=0; i<outstandingLimit; i++)
|
|
outstanding.push_back( Void() );
|
|
}
|
|
|
|
Future<Void> onSync() { // Future is set when all writes completed before the call to onSync are complete
|
|
if (outstanding.size() <= outstandingLimit)
|
|
outstanding.push_back( waitAndSync(this) );
|
|
return outstanding.back();
|
|
}
|
|
|
|
private:
|
|
int outstandingLimit;
|
|
Deque<Future<Void>> outstanding;
|
|
Reference<IAsyncFile> file;
|
|
|
|
ACTOR static Future<Void> waitAndSync(SyncQueue* self) {
|
|
Void _ = wait( self->outstanding.front() );
|
|
self->outstanding.pop_front();
|
|
Void _ = wait( self->file->sync() );
|
|
return Void();
|
|
}
|
|
};
|
|
|
|
class RawDiskQueue_TwoFiles {
|
|
public:
|
|
RawDiskQueue_TwoFiles( std::string basename, UID dbgid )
|
|
: basename(basename), onError(delayed(error.getFuture())), onStopped(stopped.getFuture()),
|
|
readingFile(-1), readingPage(-1), writingPos(-1), dbgid(dbgid),
|
|
dbg_file0BeginSeq(0), fileExtensionBytes(10<<20), readingBuffer( dbgid ),
|
|
readyToPush(Void())
|
|
{
|
|
if(BUGGIFY)
|
|
fileExtensionBytes = 8<<10;
|
|
files[0].dbgFilename = filename(0);
|
|
files[1].dbgFilename = filename(1);
|
|
stallCount.init(LiteralStringRef("RawDiskQueue.StallCount"));
|
|
}
|
|
|
|
Future<Void> pushAndCommit( StringRef pageData, StringBuffer* pageMem, uint64_t poppedPages ) {
|
|
return pushAndCommit( this, pageData, pageMem, poppedPages );
|
|
}
|
|
|
|
void stall() {
|
|
stallCount++;
|
|
readyToPush = lastCommit;
|
|
}
|
|
|
|
Future<Standalone<StringRef>> readFirstAndLastPages( compare_pages compare ) { return readFirstAndLastPages(this,compare); }
|
|
|
|
void setStartPage( int file, int64_t page ) {
|
|
TraceEvent("RDQSetStart", dbgid).detail("f",file).detail("p",page).detail("file0name", files[0].dbgFilename);
|
|
readingFile = file;
|
|
readingPage = page;
|
|
}
|
|
|
|
Future<Void> setPoppedPage( int file, int64_t page, int64_t debugSeq ) { return setPoppedPage(this, file, page, debugSeq); }
|
|
|
|
Future<Standalone<StringRef>> readNextPage() { return readNextPage(this); }
|
|
Future<Void> truncateBeforeLastReadPage() { return truncateBeforeLastReadPage(this); }
|
|
|
|
Future<Void> getError() { return onError; }
|
|
Future<Void> onClosed() { return onStopped; }
|
|
void dispose() { shutdown(this, true); }
|
|
void close() { shutdown(this, false); }
|
|
|
|
StorageBytes getStorageBytes() {
|
|
int64_t free;
|
|
int64_t total;
|
|
|
|
g_network->getDiskBytes(parentDirectory(basename), free, total);
|
|
|
|
return StorageBytes(free, total, files[0].size + files[1].size, free); // TODO: we could potentially do better in the available field by accounting for the unused pages at the end of the file
|
|
}
|
|
|
|
//private:
|
|
struct Page { uint8_t data[_PAGE_SIZE]; };
|
|
|
|
struct File {
|
|
Reference<IAsyncFile> f;
|
|
int64_t size; // always a multiple of _PAGE_SIZE, even if the physical file isn't for some reason
|
|
int64_t popped;
|
|
std::string dbgFilename;
|
|
Reference<SyncQueue> syncQueue;
|
|
|
|
File() : size(-1), popped(-1) {}
|
|
|
|
void setFile(Reference<IAsyncFile> f) {
|
|
this->f = f;
|
|
this->syncQueue = Reference<SyncQueue>( new SyncQueue(1, f) );
|
|
}
|
|
};
|
|
File files[2]; // After readFirstAndLastPages(), files[0] is logically before files[1] (pushes are always into files[1])
|
|
|
|
std::string basename;
|
|
std::string filename(int i) const { return basename + format("%d.fdq", i); }
|
|
|
|
UID dbgid;
|
|
int64_t dbg_file0BeginSeq;
|
|
|
|
Promise<Void> error, stopped;
|
|
Future<Void> onError, onStopped;
|
|
|
|
Future<Void> readyToPush;
|
|
AndFuture lastCommit;
|
|
|
|
StringBuffer readingBuffer; // Pages that have been read and not yet returned
|
|
int readingFile; // i if the next page after readingBuffer should be read from files[i], 2 if recovery is complete
|
|
int64_t readingPage; // Page within readingFile that is the next page after readingBuffer
|
|
|
|
int64_t writingPos; // Position within files[1] that will be next written
|
|
|
|
int64_t fileExtensionBytes;
|
|
|
|
AsyncMap<bool,int> recoveryActorCount;
|
|
|
|
Int64MetricHandle stallCount;
|
|
|
|
struct TrackMe : NonCopyable {
|
|
RawDiskQueue_TwoFiles* self;
|
|
TrackMe( RawDiskQueue_TwoFiles* self ) : self(self) {
|
|
self->recoveryActorCount.set(false, self->recoveryActorCount.get(false)+1);
|
|
}
|
|
~TrackMe() {
|
|
self->recoveryActorCount.set(false, self->recoveryActorCount.get(false)-1);
|
|
}
|
|
};
|
|
|
|
Future<Void> truncateFile(int file, int64_t pos) { return truncateFile(this, file, pos); }
|
|
|
|
Future<Void> push(StringRef pageData, vector<Reference<SyncQueue>>& toSync) {
|
|
// Write the given data to the queue files, swapping or extending them if necessary.
|
|
// Don't do any syncs, but push the modified file(s) onto toSync.
|
|
ASSERT( readingFile == 2 );
|
|
ASSERT( pageData.size() % _PAGE_SIZE == 0 );
|
|
ASSERT( int64_t(pageData.begin()) % _PAGE_SIZE == 0 );
|
|
ASSERT( writingPos % _PAGE_SIZE == 0 );
|
|
ASSERT( files[0].size % _PAGE_SIZE == 0 && files[1].size % _PAGE_SIZE == 0 );
|
|
|
|
vector<Future<Void>> waitfor;
|
|
|
|
if (pageData.size() + writingPos > files[1].size) {
|
|
if ( files[0].popped == files[0].size ) {
|
|
// Finish files[1] and swap
|
|
int p = files[1].size - writingPos;
|
|
if(p > 0) {
|
|
toSync.push_back( files[1].syncQueue );
|
|
/*TraceEvent("RDQWriteAndSwap", this->dbgid).detail("File1name", files[1].dbgFilename).detail("File1size", files[1].size)
|
|
.detail("writingPos", writingPos).detail("writingBytes", p);*/
|
|
waitfor.push_back( files[1].f->write( pageData.begin(), p, writingPos ) );
|
|
pageData = pageData.substr( p );
|
|
}
|
|
|
|
dbg_file0BeginSeq += files[0].size;
|
|
std::swap(files[0], files[1]);
|
|
files[1].popped = 0;
|
|
writingPos = 0;
|
|
} else {
|
|
// Extend files[1] to accomodate the new write and about 10MB or 2x current size for future writes.
|
|
/*TraceEvent("RDQExtend", this->dbgid).detail("File1name", files[1].dbgFilename).detail("File1size", files[1].size)
|
|
.detail("extensionBytes", fileExtensionBytes);*/
|
|
int64_t minExtension = pageData.size() + writingPos - files[1].size;
|
|
files[1].size += std::min(std::max(fileExtensionBytes, minExtension), files[0].size+files[1].size+minExtension);
|
|
waitfor.push_back( files[1].f->truncate( files[1].size ) );
|
|
}
|
|
}
|
|
|
|
/*TraceEvent("RDQWrite", this->dbgid).detail("File1name", files[1].dbgFilename).detail("File1size", files[1].size)
|
|
.detail("writingPos", writingPos).detail("writingBytes", pageData.size());*/
|
|
files[1].size = std::max( files[1].size, writingPos + pageData.size() );
|
|
toSync.push_back( files[1].syncQueue );
|
|
waitfor.push_back( files[1].f->write( pageData.begin(), pageData.size(), writingPos ) );
|
|
writingPos += pageData.size();
|
|
|
|
return waitForAll(waitfor);
|
|
}
|
|
|
|
ACTOR static UNCANCELLABLE Future<Void> pushAndCommit(RawDiskQueue_TwoFiles* self, StringRef pageData, StringBuffer* pageMem, uint64_t poppedPages) {
|
|
state Promise<Void> pushing, committed;
|
|
state Promise<Void> errorPromise = self->error;
|
|
state std::string filename = self->files[0].dbgFilename;
|
|
state UID dbgid = self->dbgid;
|
|
state vector<Reference<SyncQueue>> syncFiles;
|
|
try {
|
|
// pushing might need to wait for previous pushes to start (to maintain order) or for
|
|
// a previous commit to finish if stall() was called
|
|
Future<Void> ready = self->readyToPush;
|
|
self->readyToPush = pushing.getFuture();
|
|
self->lastCommit.add( committed.getFuture() );
|
|
|
|
Void _ = wait( ready );
|
|
|
|
TEST( pageData.size() > sizeof(Page) ); // push more than one page of data
|
|
|
|
Future<Void> pushed = self->push( pageData, syncFiles );
|
|
pushing.send(Void());
|
|
ASSERT( syncFiles.size() >= 1 && syncFiles.size() <= 2 );
|
|
TEST(2==syncFiles.size()); // push spans both files
|
|
Void _ = wait( pushed );
|
|
|
|
delete pageMem;
|
|
pageMem = 0;
|
|
|
|
Future<Void> sync = syncFiles[0]->onSync();
|
|
for(int i=1; i<syncFiles.size(); i++) sync = sync && syncFiles[i]->onSync();
|
|
Void _ = wait( sync );
|
|
|
|
self->updatePopped( poppedPages*sizeof(Page) );
|
|
|
|
/*TraceEvent("RDQCommitEnd", self->dbgid).detail("DeltaPopped", poppedPages*sizeof(Page)).detail("PoppedCommitted", self->dbg_file0BeginSeq + self->files[0].popped + self->files[1].popped)
|
|
.detail("File0Size", self->files[0].size).detail("File1Size", self->files[1].size)
|
|
.detail("File0Name", self->files[0].dbgFilename).detail("SyncedFiles", syncFiles.size());*/
|
|
|
|
if(g_random->random01() < 0.01) {
|
|
//occasionally delete all the ready future in the AndFuture
|
|
self->lastCommit.cleanup();
|
|
}
|
|
committed.send(Void());
|
|
} catch (Error& e) {
|
|
delete pageMem;
|
|
TEST(true); // push error
|
|
TEST(2==syncFiles.size()); // push spanning both files error
|
|
TraceEvent(SevError, "RDQ_pushAndCommit_Error", dbgid).detail("InitialFilename0", filename).error(e, true);
|
|
|
|
if (errorPromise.canBeSet()) errorPromise.sendError(e);
|
|
if (pushing.canBeSet()) pushing.sendError(e);
|
|
if (committed.canBeSet()) committed.sendError(e);
|
|
|
|
throw e;
|
|
}
|
|
return Void();
|
|
}
|
|
|
|
void updatePopped( int64_t popped ) {
|
|
int64_t pop0 = std::min(popped, files[0].size - files[0].popped);
|
|
files[0].popped += pop0;
|
|
files[1].popped += popped - pop0;
|
|
}
|
|
|
|
|
|
ACTOR static Future<Void> setPoppedPage( RawDiskQueue_TwoFiles *self, int file, int64_t page, int64_t debugSeq ) {
|
|
self->files[file].popped = page*sizeof(Page);
|
|
if (file) self->files[0].popped = self->files[0].size;
|
|
else self->files[1].popped = 0;
|
|
self->dbg_file0BeginSeq = debugSeq - self->files[1].popped - self->files[0].popped;
|
|
|
|
//If we are starting in file 1, we truncate file 0 in case it has been corrupted.
|
|
// In particular, we are trying to avoid a dropped or corrupted write to the first page of file 0 causing it to be sequenced before file 1,
|
|
// when in fact it contains many pages that follow file 1. These ok pages may be incorrectly read if the machine dies after overwritting the
|
|
// first page of file 0 and is then recovered
|
|
if(file == 1)
|
|
Void _ = wait(self->truncateFile(self, 0, 0));
|
|
|
|
return Void();
|
|
}
|
|
|
|
ACTOR static Future<Void> openFiles( RawDiskQueue_TwoFiles* self ) {
|
|
state vector<Future<Reference<IAsyncFile>>> fs;
|
|
for(int i=0; i<2; i++)
|
|
fs.push_back( IAsyncFileSystem::filesystem()->open( self->filename(i), IAsyncFile::OPEN_READWRITE | IAsyncFile::OPEN_UNCACHED | IAsyncFile::OPEN_UNBUFFERED | IAsyncFile::OPEN_LOCK, 0 ) );
|
|
Void _ = wait( waitForAllReady(fs) );
|
|
|
|
// Treatment of errors here is important. If only one of the two files is present
|
|
// (due to a power failure during creation or deletion, or administrative error) we don't want to
|
|
// open the queue!
|
|
|
|
if (!fs[0].isError() && !fs[1].isError()) {
|
|
// Both files were opened OK: success
|
|
} else if ( fs[0].isError() && fs[0].getError().code() == error_code_file_not_found &&
|
|
fs[1].isError() && fs[1].getError().code() == error_code_file_not_found )
|
|
{
|
|
// Neither file was found: we can create a new queue
|
|
// OPEN_ATOMIC_WRITE_AND_CREATE defers creation (using a .part file) until the calls to sync() below
|
|
TraceEvent("DiskQueueCreate").detail("File0", self->filename(0));
|
|
for(int i=0; i<2; i++)
|
|
fs[i] = IAsyncFileSystem::filesystem()->open( self->filename(i), IAsyncFile::OPEN_ATOMIC_WRITE_AND_CREATE | IAsyncFile::OPEN_CREATE | IAsyncFile::OPEN_READWRITE | IAsyncFile::OPEN_UNCACHED | IAsyncFile::OPEN_UNBUFFERED | IAsyncFile::OPEN_LOCK, 0600 );
|
|
|
|
// Any error here is fatal
|
|
Void _ = wait( waitForAll(fs) );
|
|
|
|
// sync on each file to actually create it will be done below
|
|
} else {
|
|
// One file had a more serious error or one file is present and the other is not. Die.
|
|
if (!fs[0].isError() || (fs[1].isError() && fs[1].getError().code() != error_code_file_not_found))
|
|
throw fs[1].getError();
|
|
else
|
|
throw fs[0].getError();
|
|
}
|
|
|
|
// fsync both files. This is necessary to trigger atomic file creation in the creation case above.
|
|
// It also permits the recovery code to assume that whatever it reads is durable. Otherwise a prior
|
|
// process could have written (but not synchronized) data to the file which we will read but which
|
|
// might not survive a reboot. The recovery code assumes otherwise and could corrupt the disk.
|
|
vector<Future<Void>> syncs;
|
|
for(int i=0; i<fs.size(); i++)
|
|
syncs.push_back( fs[i].get()->sync() );
|
|
Void _ = wait(waitForAll(syncs));
|
|
|
|
// Successfully opened or created; fill in self->files[]
|
|
for(int i=0; i<2; i++)
|
|
self->files[i].setFile(fs[i].get());
|
|
|
|
return Void();
|
|
}
|
|
|
|
ACTOR static void shutdown( RawDiskQueue_TwoFiles* self, bool deleteFiles ) {
|
|
// Wait for all reads and writes on the file, and all actors referencing self, to be finished
|
|
state Error error = success();
|
|
try {
|
|
ErrorOr<Void> _ = wait(errorOr(self->lastCommit.getFuture()));
|
|
while (self->recoveryActorCount.get(false))
|
|
Void _ = wait( self->recoveryActorCount.onChange(false) );
|
|
|
|
for(int i=0; i<2; i++)
|
|
self->files[i].f.clear();
|
|
|
|
if (deleteFiles) {
|
|
TraceEvent("DiskQueueShutdownDeleting", self->dbgid)
|
|
.detail("File0", self->filename(0))
|
|
.detail("File1", self->filename(1));
|
|
Void _ = wait( IAsyncFile::incrementalDelete( self->filename(0), false ) );
|
|
Void _ = wait( IAsyncFile::incrementalDelete( self->filename(1), true ) );
|
|
}
|
|
TraceEvent("DiskQueueShutdownComplete", self->dbgid)
|
|
.detail("DeleteFiles", deleteFiles)
|
|
.detail("File0", self->filename(0));
|
|
} catch( Error &e ) {
|
|
TraceEvent(SevError, "DiskQueueShutdownError", self->dbgid)
|
|
.detail("Reason", e.code() == error_code_platform_error ? "could not delete database" : "unknown")
|
|
.error(e,true);
|
|
error = e;
|
|
}
|
|
|
|
if( error.code() != error_code_actor_cancelled ) {
|
|
if (self->stopped.canBeSet()) self->stopped.send(Void());
|
|
if (self->error.canBeSet()) self->error.send(Never());
|
|
delete self;
|
|
}
|
|
}
|
|
|
|
ACTOR static UNCANCELLABLE Future<Standalone<StringRef>> readFirstAndLastPages(RawDiskQueue_TwoFiles* self, compare_pages compare) {
|
|
state TrackMe trackMe(self);
|
|
state StringBuffer result( self->dbgid );
|
|
|
|
try {
|
|
result.alignReserve( sizeof(Page), sizeof(Page)*3 );
|
|
state Page* firstPage = (Page*)result.append(sizeof(Page)*3);
|
|
|
|
// Open both files or create both files
|
|
Void _ = wait( openFiles(self) );
|
|
|
|
// Get the file sizes
|
|
vector<Future<int64_t>> fsize;
|
|
for(int i=0; i<2; i++)
|
|
fsize.push_back( self->files[i].f->size() );
|
|
vector<int64_t> file_sizes = wait( getAll(fsize) );
|
|
for(int i=0; i<2; i++) {
|
|
// SOMEDAY: If the file size is not a multiple of page size, it may never be shortened. Change this?
|
|
self->files[i].size = file_sizes[i] - file_sizes[i] % sizeof(Page);
|
|
ASSERT( self->files[i].size % sizeof(Page) == 0 );
|
|
}
|
|
|
|
// Read the first pages
|
|
memset(firstPage, 0, sizeof(Page)*2);
|
|
vector<Future<int>> reads;
|
|
for(int i=0; i<2; i++)
|
|
if( self->files[i].size > 0)
|
|
reads.push_back( self->files[i].f->read( &firstPage[i], sizeof(Page), 0 ) );
|
|
Void _ = wait( waitForAll(reads) );
|
|
|
|
// Determine which file comes first
|
|
if ( compare( &firstPage[1], &firstPage[0] ) ) {
|
|
std::swap( firstPage[0], firstPage[1] );
|
|
std::swap( self->files[0], self->files[1] );
|
|
}
|
|
|
|
if ( !compare( &firstPage[1], &firstPage[1] ) ) {
|
|
// Both files are invalid... the queue is empty!
|
|
// Begin pushing at the beginning of files[1]
|
|
|
|
//Truncate both files, since perhaps only the first pages are corrupted. This avoids cases where overwritting the first page and then terminating makes
|
|
//subsequent pages valid upon recovery.
|
|
vector<Future<Void>> truncates;
|
|
for(int i = 0; i < 2; ++i)
|
|
if(self->files[i].size > 0)
|
|
truncates.push_back(self->truncateFile(self, i, 0));
|
|
|
|
Void _ = wait(waitForAll(truncates));
|
|
|
|
|
|
self->files[0].popped = self->files[0].size;
|
|
self->files[1].popped = 0;
|
|
self->writingPos = 0;
|
|
self->readingFile = 2;
|
|
return Standalone<StringRef>();
|
|
}
|
|
|
|
// A page in files[1] is "valid" iff compare(&firstPage[1], page)
|
|
// Binary search to find a page in files[1] that is "valid" but the next page is not valid
|
|
// Invariant: the page at begin is valid, and the page at end is invalid
|
|
state int64_t begin = 0;
|
|
state int64_t end = self->files[1].size/sizeof(Page);
|
|
state Page *middlePage = &firstPage[2];
|
|
while ( begin + 1 != end ) {
|
|
state int64_t middle = (begin+end)/2;
|
|
ASSERT( middle > begin && middle < end ); // So the loop always changes begin or end
|
|
|
|
int len = wait( self->files[1].f->read( middlePage, sizeof(Page), middle*sizeof(Page) ) );
|
|
ASSERT( len == sizeof(Page) );
|
|
|
|
bool middleValid = compare( &firstPage[1], middlePage );
|
|
|
|
TraceEvent("RDQBS", self->dbgid).detail("b", begin).detail("e", end).detail("m", middle).detail("v", middleValid).detail("file0name", self->files[0].dbgFilename);
|
|
|
|
if (middleValid)
|
|
begin = middle;
|
|
else
|
|
end = middle;
|
|
}
|
|
// Now by the invariant and the loop condition, begin is a valid page and begin+1 is an invalid page
|
|
// Check that begin+1 is invalid
|
|
int len = wait( self->files[1].f->read( &firstPage[2], sizeof(Page), (begin+1)*sizeof(Page) ) );
|
|
ASSERT( !(len == sizeof(Page) && compare( &firstPage[1], &firstPage[2] )) );
|
|
|
|
// Read it
|
|
int len = wait( self->files[1].f->read( &firstPage[2], sizeof(Page), begin*sizeof(Page) ) );
|
|
ASSERT( len == sizeof(Page) && compare( &firstPage[1], &firstPage[2] ) );
|
|
|
|
TraceEvent("RDQEndFound", self->dbgid).detail("File0Name", self->files[0].dbgFilename).detail("Pos", begin).detail("FileSize", self->files[1].size);
|
|
|
|
return result.str;
|
|
} catch (Error& e) {
|
|
bool ok = e.code() == error_code_file_not_found;
|
|
TraceEvent(ok ? SevInfo : SevError, "RDQ_rfl_Error", self->dbgid).detail("file0name", self->files[0].dbgFilename).error(e, true);
|
|
if (!self->error.isSet()) self->error.sendError(e);
|
|
throw;
|
|
}
|
|
}
|
|
|
|
Future<int> fillReadingBuffer() {
|
|
// If we're right at the end of a file...
|
|
if ( readingPage*sizeof(Page) >= (size_t)files[readingFile].size ) {
|
|
readingFile++;
|
|
readingPage = 0;
|
|
if (readingFile>=2) {
|
|
// Recovery complete
|
|
readingBuffer.clear();
|
|
writingPos = files[1].size;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
// Read up to 1MB into readingBuffer
|
|
int len = std::min<int64_t>( (files[readingFile].size/sizeof(Page) - readingPage)*sizeof(Page), BUGGIFY_WITH_PROB(1.0) ? sizeof(Page)*g_random->randomInt(1,4) : (1<<20) );
|
|
readingBuffer.clear();
|
|
readingBuffer.alignReserve( sizeof(Page), len );
|
|
void* p = readingBuffer.append(len);
|
|
|
|
auto pos = readingPage * sizeof(Page);
|
|
readingPage += len / sizeof(Page);
|
|
ASSERT( int64_t(p) % sizeof(Page) == 0 );
|
|
return files[readingFile].f->read( p, len, pos );
|
|
}
|
|
|
|
ACTOR static UNCANCELLABLE Future<Standalone<StringRef>> readNextPage(RawDiskQueue_TwoFiles* self) {
|
|
state TrackMe trackMe(self);
|
|
|
|
try {
|
|
ASSERT( self->readingFile < 2 );
|
|
ASSERT( self->files[0].f && self->files[1].f );
|
|
|
|
|
|
if (!self->readingBuffer.size()) {
|
|
state Future<Void> f = Void();
|
|
//if (BUGGIFY) f = delay( g_random->random01() * 0.1 );
|
|
|
|
int read = wait( self->fillReadingBuffer() );
|
|
ASSERT( read == self->readingBuffer.size() );
|
|
|
|
Void _ = wait(f);
|
|
}
|
|
if (!self->readingBuffer.size()) return Standalone<StringRef>();
|
|
|
|
ASSERT( self->readingBuffer.size() >= sizeof(Page) );
|
|
Standalone<StringRef> result = self->readingBuffer.pop_front( sizeof(Page) );
|
|
return result;
|
|
} catch (Error& e) {
|
|
TEST(true); // Read next page error
|
|
TraceEvent(SevError, "RDQ_rnp_Error", self->dbgid).detail("file0name", self->files[0].dbgFilename).error(e, true);
|
|
if (!self->error.isSet()) self->error.sendError(e);
|
|
throw;
|
|
}
|
|
}
|
|
|
|
ACTOR static UNCANCELLABLE Future<Void> truncateFile(RawDiskQueue_TwoFiles* self, int file, int64_t pos) {
|
|
state TrackMe trackMe(self);
|
|
state StringBuffer zeros( self->dbgid );
|
|
|
|
TraceEvent("DQTruncateFile", self->dbgid).detail("File", file).detail("Pos", pos).detail("File0Name", self->files[0].dbgFilename);
|
|
|
|
zeros.alignReserve( sizeof(Page), 1<<20 );
|
|
memset( zeros.append(1<<20), 0, 1<<20 );
|
|
|
|
while(pos < self->files[file].size) {
|
|
state int len = std::min<int64_t>(zeros.size(), self->files[file].size-pos);
|
|
Void _ = wait( self->files[file].f->write( zeros.str.begin(), len, pos ) );
|
|
pos += len;
|
|
}
|
|
|
|
Void _ = wait(self->files[file].syncQueue->onSync());
|
|
return Void();
|
|
}
|
|
|
|
ACTOR static Future<Void> truncateBeforeLastReadPage( RawDiskQueue_TwoFiles* self ) {
|
|
try {
|
|
state int file = self->readingFile;
|
|
state int64_t pos = (self->readingPage - self->readingBuffer.size()/sizeof(Page) - 1) * sizeof(Page);
|
|
state vector<Future<Void>> commits;
|
|
state bool swap = file==0;
|
|
|
|
TEST( file==0 ); // truncate before last read page on file 0
|
|
TEST( file==1 && pos != self->files[1].size ); // truncate before last read page on file 1
|
|
|
|
self->readingFile = 2;
|
|
self->readingBuffer.clear();
|
|
self->writingPos = pos;
|
|
|
|
while (file < 2) {
|
|
commits.push_back(self->truncateFile(self, file, pos));
|
|
file++;
|
|
pos = 0;
|
|
}
|
|
|
|
Void _ = wait( waitForAll(commits) );
|
|
|
|
if (swap) {
|
|
std::swap(self->files[0], self->files[1]);
|
|
self->files[0].popped = self->files[0].size;
|
|
}
|
|
|
|
return Void();
|
|
} catch (Error& e) {
|
|
TraceEvent(SevError, "RDQ_tblrp_Error", self->dbgid).detail("file0name", self->files[0].dbgFilename).error(e);
|
|
if (!self->error.isSet()) self->error.sendError(e);
|
|
throw;
|
|
}
|
|
}
|
|
};
|
|
|
|
class DiskQueue : public IDiskQueue {
|
|
public:
|
|
DiskQueue( std::string basename, UID dbgid )
|
|
: rawQueue( new RawDiskQueue_TwoFiles(basename, dbgid) ), dbgid(dbgid), anyPopped(false), nextPageSeq(0), poppedSeq(0), lastPoppedSeq(0),
|
|
nextReadLocation(-1), readBufPage(NULL), readBufPos(0), pushed_page_buffer(NULL), recovered(false), lastCommittedSeq(0), warnAlwaysForMemory(true)
|
|
{
|
|
}
|
|
|
|
virtual location push( StringRef contents ) {
|
|
ASSERT( recovered );
|
|
uint8_t const* begin = contents.begin();
|
|
uint8_t const* end = contents.end();
|
|
TEST( contents.size() && pushedPageCount() ); // More than one push between commits
|
|
TEST( contents.size()>=4 && pushedPageCount() && backPage().remainingCapacity()<4 ); // Push right at the end of a page, possibly splitting size
|
|
while (begin != end) {
|
|
if (!pushedPageCount() || !backPage().remainingCapacity()) addEmptyPage();
|
|
|
|
auto &p = backPage();
|
|
int s = std::min<int>( p.remainingCapacity(), end-begin );
|
|
memcpy( p.payload + p.payloadSize, begin, s );
|
|
p.payloadSize += s;
|
|
begin += s;
|
|
}
|
|
return endLocation();
|
|
}
|
|
virtual void pop( location upTo ) {
|
|
ASSERT( !upTo.hi );
|
|
ASSERT( !recovered || upTo.lo <= endLocation() );
|
|
|
|
// The following ASSERT is NOT part of the intended contract of IDiskQueue, but alerts the user to a known bug where popping
|
|
// into uncommitted pages can cause a durability failure.
|
|
// FIXME: Remove this ASSERT when popping into uncommitted pages is fixed
|
|
if( upTo.lo > lastCommittedSeq ) {
|
|
TraceEvent(SevError, "DQPopUncommittedData", dbgid)
|
|
.detail("UpTo", upTo)
|
|
.detail("LastCommittedSeq", lastCommittedSeq)
|
|
.detail("file0name", rawQueue->files[0].dbgFilename);
|
|
}
|
|
if (upTo.lo > poppedSeq) {
|
|
poppedSeq = upTo.lo;
|
|
anyPopped = true;
|
|
}
|
|
}
|
|
|
|
int getMaxPayload() {
|
|
return Page::maxPayload;
|
|
}
|
|
|
|
virtual int getCommitOverhead() {
|
|
if(!pushedPageCount()) {
|
|
if(!anyPopped)
|
|
return 0;
|
|
|
|
return Page::maxPayload;
|
|
}
|
|
else
|
|
return backPage().remainingCapacity();
|
|
}
|
|
|
|
virtual Future<Void> commit() {
|
|
ASSERT( recovered );
|
|
if (!pushedPageCount()) {
|
|
if (!anyPopped) return Void();
|
|
anyPopped = false;
|
|
addEmptyPage();
|
|
}
|
|
backPage().popped = poppedSeq;
|
|
backPage().zeroPad();
|
|
backPage().updateHash();
|
|
|
|
if( pushedPageCount() >= 8000 ) {
|
|
TraceEvent( warnAlwaysForMemory ? SevWarnAlways : SevWarn, "DiskQueueMemoryWarning", dbgid)
|
|
.detail("pushed_pages", pushedPageCount())
|
|
.detail("nextPageSeq", nextPageSeq)
|
|
.detail("Details", format("%d pages", pushedPageCount()))
|
|
.detail("file0name", rawQueue->files[0].dbgFilename)
|
|
.suppressFor(1.0);
|
|
if(g_network->isSimulated())
|
|
warnAlwaysForMemory = false;
|
|
}
|
|
|
|
/*TraceEvent("DQCommit", dbgid).detail("Pages", pushedPageCount()).detail("lastPoppedSeq", lastPoppedSeq).detail("poppedSeq", poppedSeq).detail("nextPageSeq", nextPageSeq)
|
|
.detail("RawFile0Size", rawQueue->files[0].size).detail("RawFile1Size", rawQueue->files[1].size).detail("writingPos", rawQueue->writingPos)
|
|
.detail("RawFile0Name", rawQueue->files[0].dbgFilename);*/
|
|
|
|
lastCommittedSeq = backPage().endSeq();
|
|
auto f = rawQueue->pushAndCommit( pushed_page_buffer->ref(), pushed_page_buffer, poppedSeq/sizeof(Page) - lastPoppedSeq/sizeof(Page) );
|
|
lastPoppedSeq = poppedSeq;
|
|
pushed_page_buffer = 0;
|
|
return f;
|
|
}
|
|
void stall() {
|
|
rawQueue->stall();
|
|
}
|
|
|
|
virtual Future<Standalone<StringRef>> readNext( int bytes ) { return readNext(this, bytes); }
|
|
|
|
virtual location getNextReadLocation() { return nextReadLocation; }
|
|
|
|
virtual Future<Void> getError() { return rawQueue->getError(); }
|
|
virtual Future<Void> onClosed() { return rawQueue->onClosed(); }
|
|
virtual void dispose() {
|
|
TraceEvent("DQDestroy", dbgid).detail("lastPoppedSeq", lastPoppedSeq).detail("poppedSeq", poppedSeq).detail("nextPageSeq", nextPageSeq).detail("file0name", rawQueue->files[0].dbgFilename);
|
|
rawQueue->dispose();
|
|
delete this;
|
|
}
|
|
virtual void close() {
|
|
TraceEvent("DQClose", dbgid)
|
|
.detail("lastPoppedSeq", lastPoppedSeq)
|
|
.detail("poppedSeq", poppedSeq)
|
|
.detail("nextPageSeq", nextPageSeq)
|
|
.detail("poppedCommitted", rawQueue->dbg_file0BeginSeq + rawQueue->files[0].popped + rawQueue->files[1].popped)
|
|
.detail("file0name", rawQueue->files[0].dbgFilename);
|
|
rawQueue->close();
|
|
delete this;
|
|
}
|
|
|
|
virtual StorageBytes getStorageBytes() {
|
|
return rawQueue->getStorageBytes();
|
|
}
|
|
|
|
private:
|
|
#pragma pack(push, 1)
|
|
struct PageHeader {
|
|
UID hash;
|
|
uint64_t seq;
|
|
uint64_t popped;
|
|
int payloadSize;
|
|
};
|
|
|
|
struct Page : PageHeader {
|
|
static const int maxPayload = _PAGE_SIZE - sizeof(PageHeader);
|
|
uint8_t payload[maxPayload];
|
|
|
|
int remainingCapacity() const { return maxPayload - payloadSize; }
|
|
uint64_t endSeq() const { return seq + sizeof(PageHeader) + payloadSize; }
|
|
void updateHash() {
|
|
// SOMEDAY: Better hash?
|
|
uint32_t part[2] = { 0x12345678, 0xbeefabcd };
|
|
hashlittle2( &seq, sizeof(Page)-sizeof(hash), &part[0], &part[1] );
|
|
hash = UID( (int64_t(part[0])<<32)+part[1], 0xfdb );
|
|
}
|
|
bool checkHash() {
|
|
UID h = hash;
|
|
updateHash();
|
|
if (h != hash) { std::swap(h, hash); return false; }
|
|
return true;
|
|
}
|
|
void zeroPad() {
|
|
memset( payload+payloadSize, 0, maxPayload-payloadSize );
|
|
}
|
|
};
|
|
#pragma pack(pop)
|
|
|
|
loc_t endLocation() const { return pushedPageCount() ? backPage().endSeq() : nextPageSeq; }
|
|
|
|
void addEmptyPage() {
|
|
if (pushedPageCount()) {
|
|
backPage().updateHash();
|
|
ASSERT( backPage().payloadSize == Page::maxPayload );
|
|
}
|
|
|
|
//pushed_pages.resize( pushed_pages.arena(), pushed_pages.size()+1 );
|
|
if (!pushed_page_buffer) pushed_page_buffer = new StringBuffer( dbgid );
|
|
pushed_page_buffer->alignReserve( sizeof(Page), pushed_page_buffer->size() + sizeof(Page) );
|
|
pushed_page_buffer->append( sizeof(Page) );
|
|
|
|
ASSERT( nextPageSeq%sizeof(Page)==0 );
|
|
|
|
auto& p = backPage();
|
|
memset(&p, 0, sizeof(Page)); // FIXME: unnecessary?
|
|
p.payloadSize = 0;
|
|
p.seq = nextPageSeq;
|
|
nextPageSeq += sizeof(Page);
|
|
p.popped = poppedSeq;
|
|
|
|
if (pushedPageCount() == 8000) {
|
|
TraceEvent("DiskQueueHighPageCount", dbgid)
|
|
.detail("pushed_pages", pushedPageCount())
|
|
.detail("nextPageSeq", nextPageSeq)
|
|
.detail("file0name", rawQueue->files[0].dbgFilename);
|
|
}
|
|
}
|
|
|
|
void readFromBuffer( StringBuffer& result, int& bytes ) {
|
|
// extract up to bytes from readBufPage into result
|
|
int len = std::min( readBufPage->payloadSize - readBufPos, bytes );
|
|
if (len<=0) return;
|
|
|
|
result.append( StringRef(readBufPage->payload+readBufPos, len) );
|
|
|
|
readBufPos += len;
|
|
bytes -= len;
|
|
nextReadLocation += len;
|
|
}
|
|
|
|
ACTOR static Future<Standalone<StringRef>> readNext( DiskQueue *self, int bytes ) {
|
|
state StringBuffer result( self->dbgid );
|
|
ASSERT(bytes >= 0);
|
|
result.clearReserve(bytes);
|
|
|
|
ASSERT( !self->recovered );
|
|
|
|
if (self->nextReadLocation < 0) {
|
|
bool nonempty = wait( findStart(self) );
|
|
if (!nonempty) {
|
|
// The constructor has already put everything in the right state for an empty queue
|
|
self->recovered = true;
|
|
ASSERT( self->poppedSeq <= self->endLocation() );
|
|
|
|
//The next read location isn't necessarily the end of the last commit, but this is sufficient for helping us check an ASSERTion
|
|
self->lastCommittedSeq = self->nextReadLocation;
|
|
|
|
return Standalone<StringRef>();
|
|
}
|
|
self->readBufPos = self->nextReadLocation % sizeof(Page) - sizeof(PageHeader);
|
|
if (self->readBufPos < 0) { self->nextReadLocation -= self->readBufPos; self->readBufPos = 0; }
|
|
TraceEvent("DQRecStart", self->dbgid).detail("readBufPos", self->readBufPos).detail("nextReadLoc", self->nextReadLocation).detail("file0name", self->rawQueue->files[0].dbgFilename);
|
|
}
|
|
|
|
loop {
|
|
if (self->readBufPage) {
|
|
self->readFromBuffer( result, bytes );
|
|
// if done, return
|
|
if (!bytes) return result.str;
|
|
ASSERT( self->readBufPos == self->readBufPage->payloadSize );
|
|
self->readBufPage = 0;
|
|
self->nextReadLocation += sizeof(Page) - self->readBufPos;
|
|
self->readBufPos = 0;
|
|
}
|
|
|
|
Standalone<StringRef> page = wait( self->rawQueue->readNextPage() );
|
|
if (!page.size()) {
|
|
TraceEvent("DQRecEOF", self->dbgid).detail("nextReadLocation", self->nextReadLocation).detail("file0name", self->rawQueue->files[0].dbgFilename);
|
|
break;
|
|
}
|
|
ASSERT( page.size() == sizeof(Page) );
|
|
|
|
self->readBufArena = page.arena();
|
|
self->readBufPage = (Page*)page.begin();
|
|
if (!self->readBufPage->checkHash() || self->readBufPage->seq < self->nextReadLocation/sizeof(Page)*sizeof(Page)) {
|
|
TraceEvent("DQRecInvalidPage", self->dbgid).detail("nextReadLocation", self->nextReadLocation).detail("hashCheck", self->readBufPage->checkHash())
|
|
.detail("seq", self->readBufPage->seq).detail("expect", self->nextReadLocation/sizeof(Page)*sizeof(Page)).detail("file0name", self->rawQueue->files[0].dbgFilename);
|
|
Void _ = wait( self->rawQueue->truncateBeforeLastReadPage() );
|
|
break;
|
|
}
|
|
//TraceEvent("DQRecPage", self->dbgid).detail("nextReadLoc", self->nextReadLocation).detail("Seq", self->readBufPage->seq).detail("Pop", self->readBufPage->popped).detail("Payload", self->readBufPage->payloadSize).detail("file0name", self->rawQueue->files[0].dbgFilename);
|
|
ASSERT( self->readBufPage->seq == self->nextReadLocation/sizeof(Page)*sizeof(Page) );
|
|
self->lastPoppedSeq = self->readBufPage->popped;
|
|
}
|
|
|
|
// Recovery complete.
|
|
// The fully durable popped point is self->lastPoppedSeq; tell the raw queue that.
|
|
int f; int64_t p;
|
|
TEST( self->lastPoppedSeq/sizeof(Page) != self->poppedSeq/sizeof(Page) ); // DiskQueue: Recovery popped position not fully durable
|
|
self->findPhysicalLocation( self->lastPoppedSeq, f, p, "lastPoppedSeq" );
|
|
Void _ = wait(self->rawQueue->setPoppedPage( f, p, self->lastPoppedSeq/sizeof(Page)*sizeof(Page) ));
|
|
|
|
// Writes go at the end of our reads (but on the next page)
|
|
self->nextPageSeq = self->nextReadLocation/sizeof(Page)*sizeof(Page);
|
|
if (self->nextReadLocation % sizeof(Page) > 36) self->nextPageSeq += sizeof(Page);
|
|
|
|
TraceEvent("DQRecovered", self->dbgid).detail("lastPoppedSeq", self->lastPoppedSeq).detail("poppedSeq", self->poppedSeq).detail("nextPageSeq", self->nextPageSeq).detail("file0name", self->rawQueue->files[0].dbgFilename);
|
|
self->recovered = true;
|
|
ASSERT( self->poppedSeq <= self->endLocation() );
|
|
self->recoveryFirstPages = Standalone<StringRef>();
|
|
|
|
TEST( result.size() == 0 ); // End of queue at border between reads
|
|
TEST( result.size() != 0 ); // Partial read at end of queue
|
|
|
|
//The next read location isn't necessarily the end of the last commit, but this is sufficient for helping us check an ASSERTion
|
|
self->lastCommittedSeq = self->nextReadLocation;
|
|
|
|
return result.str;
|
|
}
|
|
|
|
ACTOR static Future<bool> findStart( DiskQueue* self ) {
|
|
Standalone<StringRef> epbuf = wait( self->rawQueue->readFirstAndLastPages( &comparePages ) );
|
|
ASSERT( epbuf.size() % sizeof(Page) == 0 );
|
|
self->recoveryFirstPages = epbuf;
|
|
|
|
if (!epbuf.size()) {
|
|
// There are no valid pages, so apparently this is a completely empty queue
|
|
self->nextReadLocation = 0;
|
|
return false;
|
|
}
|
|
|
|
int n = epbuf.size() / sizeof(Page);
|
|
Page* lastPage = (Page*)epbuf.end() - 1;
|
|
self->nextReadLocation = self->poppedSeq = lastPage->popped;
|
|
|
|
/*
|
|
state std::auto_ptr<Page> testPage(new Page);
|
|
state int fileNum;
|
|
for( fileNum=0; fileNum<2; fileNum++) {
|
|
state int sizeNum;
|
|
for( sizeNum=0; sizeNum < self->rawQueue->files[fileNum].size; sizeNum += sizeof(Page) ) {
|
|
int _ = wait( self->rawQueue->files[fileNum].f->read( testPage.get(), sizeof(Page), sizeNum ) );
|
|
TraceEvent("PageData").detail("file", self->rawQueue->files[fileNum].dbgFilename).detail("sizeNum", sizeNum).detail("seq", testPage->seq).detail("hash", testPage->checkHash()).detail("popped", testPage->popped);
|
|
}
|
|
}
|
|
*/
|
|
|
|
int file; int64_t page;
|
|
self->findPhysicalLocation( self->poppedSeq, file, page, "poppedSeq" );
|
|
self->rawQueue->setStartPage( file, page );
|
|
|
|
return true;
|
|
}
|
|
|
|
void findPhysicalLocation( loc_t loc, int& file, int64_t& page, const char* context ) {
|
|
bool ok = false;
|
|
Page*p = (Page*)recoveryFirstPages.begin();
|
|
|
|
TraceEvent(SevInfo, "FindPhysicalLocation", dbgid)
|
|
.detail("RecoveryFirstPages", recoveryFirstPages.size())
|
|
.detail("Page0Valid", p[0].checkHash())
|
|
.detail("Page0Seq", p[0].seq)
|
|
.detail("Page1Valid", p[1].checkHash())
|
|
.detail("Page1Seq", p[1].seq)
|
|
.detail("Location", loc)
|
|
.detail("Context", context)
|
|
.detail("file0name", rawQueue->files[0].dbgFilename);
|
|
|
|
for(int i=recoveryFirstPages.size() / sizeof(Page) - 2; i>=0; i--)
|
|
if ( p[i].checkHash() && p[i].seq <= (size_t)loc ) {
|
|
file = i;
|
|
page = (loc - p[i].seq)/sizeof(Page);
|
|
TraceEvent("FoundPhysicalLocation", dbgid)
|
|
.detail("PageIndex", i)
|
|
.detail("PageLocation", page)
|
|
.detail("RecoveryFirstPagesSize", recoveryFirstPages.size())
|
|
.detail("SizeofPage", sizeof(Page))
|
|
.detail("PageSequence", p[i].seq)
|
|
.detail("Location", loc)
|
|
.detail("Context", context)
|
|
.detail("file0name", rawQueue->files[0].dbgFilename);
|
|
ok = true;
|
|
break;
|
|
}
|
|
if (!ok)
|
|
TraceEvent(SevError, "DiskQueueLocationError", dbgid)
|
|
.detail("RecoveryFirstPages", recoveryFirstPages.size())
|
|
.detail("Page0Valid", p[0].checkHash())
|
|
.detail("Page0Seq", p[0].seq)
|
|
.detail("Page1Valid", p[1].checkHash())
|
|
.detail("Page1Seq", p[1].seq)
|
|
.detail("Location", loc)
|
|
.detail("Context", context)
|
|
.detail("file0name", rawQueue->files[0].dbgFilename);
|
|
ASSERT( ok );
|
|
}
|
|
|
|
// isValid(firstPage) == compare(firstPage, firstPage)
|
|
// isValid(otherPage) == compare(firstPage, otherPage)
|
|
// Swap file1, file2 if comparePages( file2.firstPage, file1.firstPage )
|
|
static bool comparePages( void* v1, void* v2 ) {
|
|
Page* p1 = (Page*)v1; Page* p2 = (Page*)v2;
|
|
return p2->checkHash() && (p2->seq >= p1->seq || !p1->checkHash());
|
|
}
|
|
|
|
RawDiskQueue_TwoFiles *rawQueue;
|
|
UID dbgid;
|
|
|
|
bool anyPopped; // pop() has been called since the most recent commit()
|
|
bool warnAlwaysForMemory;
|
|
loc_t nextPageSeq, poppedSeq;
|
|
loc_t lastPoppedSeq; // poppedSeq the last time commit was called
|
|
loc_t lastCommittedSeq;
|
|
|
|
// Buffer of pushed pages that haven't been committed. The last one (backPage()) is still mutable.
|
|
StringBuffer* pushed_page_buffer;
|
|
Page& backPage() {
|
|
ASSERT( pushedPageCount() );
|
|
return ((Page*)pushed_page_buffer->ref().end())[-1];
|
|
}
|
|
Page const& backPage() const { return ((Page*)pushed_page_buffer->ref().end())[-1]; }
|
|
int pushedPageCount() const { return pushed_page_buffer ? pushed_page_buffer->size() / sizeof(Page) : 0; }
|
|
|
|
// Recovery state
|
|
bool recovered;
|
|
loc_t nextReadLocation;
|
|
Arena readBufArena;
|
|
Page* readBufPage;
|
|
int readBufPos;
|
|
Standalone<StringRef> recoveryFirstPages;
|
|
};
|
|
|
|
//A class wrapping DiskQueue which durably allows uncommitted data to be popped
|
|
//This works by performing two commits when uncommitted data is popped:
|
|
// Commit 1 - pop only previously committed data and push new data
|
|
// Commit 2 - finish pop into uncommitted data
|
|
class DiskQueue_PopUncommitted : public IDiskQueue {
|
|
|
|
public:
|
|
DiskQueue_PopUncommitted( std::string basename, UID dbgid ) : queue(new DiskQueue(basename, dbgid)), pushed(0), popped(0), committed(0) { };
|
|
|
|
//IClosable
|
|
Future<Void> getError() { return queue->getError(); }
|
|
Future<Void> onClosed() { return queue->onClosed(); }
|
|
void dispose() { queue->dispose(); delete this; }
|
|
void close() { queue->close(); delete this; }
|
|
|
|
//IDiskQueue
|
|
Future<Standalone<StringRef>> readNext( int bytes ) { return readNext(this, bytes); }
|
|
|
|
virtual location getNextReadLocation() { return queue->getNextReadLocation(); }
|
|
|
|
virtual location push( StringRef contents ) {
|
|
pushed = queue->push(contents);
|
|
return pushed;
|
|
}
|
|
|
|
virtual void pop( location upTo ) {
|
|
popped = std::max(popped, upTo);
|
|
ASSERT_WE_THINK(committed >= popped);
|
|
queue->pop(std::min(committed, popped));
|
|
}
|
|
|
|
virtual int getCommitOverhead() {
|
|
return queue->getCommitOverhead() + (popped > committed ? queue->getMaxPayload() : 0);
|
|
}
|
|
|
|
Future<Void> commit() {
|
|
location pushLocation = pushed;
|
|
location popLocation = popped;
|
|
|
|
Future<Void> commitFuture = queue->commit();
|
|
|
|
bool updatePop = popLocation > committed;
|
|
committed = pushLocation;
|
|
|
|
if(updatePop) {
|
|
ASSERT_WE_THINK(false);
|
|
ASSERT(popLocation <= committed);
|
|
|
|
queue->stall(); // Don't permit this pipelined commit to write anything to disk until the previous commit is totally finished
|
|
pop(popLocation);
|
|
commitFuture = commitFuture && queue->commit();
|
|
}
|
|
else
|
|
TEST(true); //No uncommitted data was popped
|
|
|
|
return commitFuture;
|
|
}
|
|
|
|
virtual StorageBytes getStorageBytes() { return queue->getStorageBytes(); }
|
|
|
|
private:
|
|
DiskQueue *queue;
|
|
location pushed;
|
|
location popped;
|
|
location committed;
|
|
|
|
ACTOR static Future<Standalone<StringRef>> readNext( DiskQueue_PopUncommitted *self, int bytes ) {
|
|
Standalone<StringRef> str = wait(self->queue->readNext(bytes));
|
|
if(str.size() < bytes)
|
|
self->pushed = self->getNextReadLocation();
|
|
|
|
return str;
|
|
}
|
|
};
|
|
|
|
IDiskQueue* openDiskQueue( std::string basename, UID dbgid ) {
|
|
return new DiskQueue_PopUncommitted( basename, dbgid );
|
|
}
|