foundationdb/fdbrpc/QueueModel.h

124 lines
4.7 KiB
C
Raw Normal View History

2017-05-26 04:48:44 +08:00
/*
* QueueModel.h
*
* This source file is part of the FoundationDB open source project
*
* Copyright 2013-2018 Apple Inc. and the FoundationDB project authors
*
2017-05-26 04:48:44 +08:00
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
2017-05-26 04:48:44 +08:00
* http://www.apache.org/licenses/LICENSE-2.0
*
2017-05-26 04:48:44 +08:00
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef FLOW_QUEUEMODEL_H
#define FLOW_QUEUEMODEL_H
#pragma once
#include "flow/flow.h"
#include "fdbrpc/Smoother.h"
2017-05-26 04:48:44 +08:00
#include "flow/Knobs.h"
#include "flow/ActorCollection.h"
// The data structure used for the client-side load balancing algorithm to
// decide which storage server to read data from. Conceptually, it tracks the
// number of outstanding requests the current client sent to each storage
// server. One "QueueData" represents one storage server.
2017-05-26 04:48:44 +08:00
struct QueueData {
// The current outstanding requests sent by the local client to this storage
// server. The number is smoothed out over a continuous timeline.
2017-05-26 04:48:44 +08:00
Smoother smoothOutstanding;
// The last client perceived latency to this storage server.
2017-05-26 04:48:44 +08:00
double latency;
// Represents the "cost" of each storage request. By default, the penalty is
// 1 indicates that each outstanding request corresponds 1 outstanding
// request. However, storage server can also increase the penalty if it
// decides to ask the client to slow down sending requests to it. Penalty
// is updated after each LoadBalancedReply.
2017-05-26 04:48:44 +08:00
double penalty;
// Do not consider this storage server if the current time hasn't reach this
// time. This field is computed after each request to not repeatedly try the
// same storage server that is likely not going to return a valid result.
double failedUntil;
// If the storage server returns a "future version" error, increase above
// `failedUntil` by this amount to increase the backoff time.
double futureVersionBackoff;
// If the current time has reached this time, and this storage server still
// hasn't returned a valid result, increase above `futureVersionBackoff`
// to increase the future backoff amount.
double increaseBackoffTime;
QueueData()
: latency(0.001), penalty(1.0), smoothOutstanding(FLOW_KNOBS->QUEUE_MODEL_SMOOTHING_AMOUNT), failedUntil(0),
futureVersionBackoff(FLOW_KNOBS->FUTURE_VERSION_INITIAL_BACKOFF), increaseBackoffTime(0) {}
2017-05-26 04:48:44 +08:00
};
typedef double TimeEstimate;
class QueueModel {
public:
// Finishes the request sent to storage server with `id`.
// - latency: the measured client-side latency of the request.
// - penalty: the server side penalty sent along with the response from
// the storage server. Requires >= 1.
// - delta: Update server `id`'s queue model by substract this amount.
// This value should be the value returned by `addRequest` below.
// - clean: indicates whether the there was an error or not.
// - futureVersion: indicates whether there was "future version" error or
// not.
void endRequest(uint64_t id, double latency, double penalty, double delta, bool clean, bool futureVersion);
QueueData& getMeasurement(uint64_t id);
// Starts a new request to storage server with `id`. If the storage
// server contains a penalty, add it to the queue size, and return the
// penalty. The returned penalty should be passed as `delta` to `endRequest`
// to make `smoothOutstanding` to reflect the real storage queue size.
double addRequest(uint64_t id);
2017-05-26 04:48:44 +08:00
double secondMultiplier;
double secondBudget;
PromiseStream<Future<Void>> addActor;
2017-05-26 04:48:44 +08:00
Future<Void> laggingRequests; // requests for which a different recipient already answered
int laggingRequestCount;
QueueModel() : secondMultiplier(1.0), secondBudget(0), laggingRequestCount(0) {
laggingRequests = actorCollection(addActor.getFuture(), &laggingRequestCount);
2017-05-26 04:48:44 +08:00
}
~QueueModel() { laggingRequests.cancel(); }
2017-05-26 04:48:44 +08:00
private:
2019-03-28 11:56:44 +08:00
std::unordered_map<uint64_t, QueueData> data;
2017-05-26 04:48:44 +08:00
};
/* old queue model
class QueueModel {
public:
QueueModel() : new_index(0) {
total_time[0] = 0;
total_time[1] = 0;
}
void addMeasurement( uint64_t id, QueueDetails qd );
TimeEstimate getTimeEstimate( uint64_t id );
TimeEstimate getAverageTimeEstimate();
QueueDetails getMeasurement( uint64_t id );
void expire();
2017-05-26 04:48:44 +08:00
private:
std::map<uint64_t, QueueDetails> data[2];
double total_time[2];
int new_index; // data[new_index] is the new data
2017-05-26 04:48:44 +08:00
};
*/
#endif