mirror of
https://github.com/apple/foundationdb.git
synced 2025-05-14 01:42:37 +08:00
More blob granule operational stuff (#7783)
* More blob manager metrics * making blobrange check command work for large ranges
This commit is contained in:
parent
b9d156f0d9
commit
1cda8a2fc1
@ -99,20 +99,68 @@ ACTOR Future<Void> doBlobPurge(Database db, Key startKey, Key endKey, Optional<V
|
|||||||
return Void();
|
return Void();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
ACTOR Future<Version> checkBlobSubrange(Database db, KeyRange keyRange, Optional<Version> version) {
|
||||||
|
state Transaction tr(db);
|
||||||
|
state Version readVersionOut = invalidVersion;
|
||||||
|
loop {
|
||||||
|
try {
|
||||||
|
wait(success(tr.readBlobGranules(keyRange, 0, version, &readVersionOut)));
|
||||||
|
return readVersionOut;
|
||||||
|
} catch (Error& e) {
|
||||||
|
wait(tr.onError(e));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
ACTOR Future<Void> doBlobCheck(Database db, Key startKey, Key endKey, Optional<Version> version) {
|
ACTOR Future<Void> doBlobCheck(Database db, Key startKey, Key endKey, Optional<Version> version) {
|
||||||
state Transaction tr(db);
|
state Transaction tr(db);
|
||||||
state Version readVersionOut = invalidVersion;
|
state Version readVersionOut = invalidVersion;
|
||||||
state double elapsed = -timer_monotonic();
|
state double elapsed = -timer_monotonic();
|
||||||
|
state KeyRange range = KeyRange(KeyRangeRef(startKey, endKey));
|
||||||
|
state Standalone<VectorRef<KeyRangeRef>> allRanges;
|
||||||
loop {
|
loop {
|
||||||
try {
|
try {
|
||||||
wait(success(tr.readBlobGranules(KeyRange(KeyRangeRef(startKey, endKey)), 0, version, &readVersionOut)));
|
wait(store(allRanges, tr.getBlobGranuleRanges(range)));
|
||||||
elapsed += timer_monotonic();
|
|
||||||
break;
|
break;
|
||||||
} catch (Error& e) {
|
} catch (Error& e) {
|
||||||
wait(tr.onError(e));
|
wait(tr.onError(e));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (allRanges.empty()) {
|
||||||
|
fmt::print("ERROR: No blob ranges for [{0} - {1})\n", startKey.printable(), endKey.printable());
|
||||||
|
return Void();
|
||||||
|
}
|
||||||
|
fmt::print("Loaded {0} blob ranges to check\n", allRanges.size());
|
||||||
|
state std::vector<Future<Version>> checkParts;
|
||||||
|
// chunk up to smaller ranges than max
|
||||||
|
int maxChunkSize = 1000;
|
||||||
|
KeyRange currentChunk;
|
||||||
|
int currentChunkSize = 0;
|
||||||
|
for (auto& it : allRanges) {
|
||||||
|
if (currentChunkSize == maxChunkSize) {
|
||||||
|
checkParts.push_back(checkBlobSubrange(db, currentChunk, version));
|
||||||
|
currentChunkSize = 0;
|
||||||
|
}
|
||||||
|
if (currentChunkSize == 0) {
|
||||||
|
currentChunk = it;
|
||||||
|
} else if (it.begin != currentChunk.end) {
|
||||||
|
fmt::print("ERROR: Blobrange check failed, gap in blob ranges from [{0} - {1})\n",
|
||||||
|
currentChunk.end.printable(),
|
||||||
|
it.begin.printable());
|
||||||
|
return Void();
|
||||||
|
} else {
|
||||||
|
currentChunk = KeyRangeRef(currentChunk.begin, it.end);
|
||||||
|
}
|
||||||
|
currentChunkSize++;
|
||||||
|
}
|
||||||
|
checkParts.push_back(checkBlobSubrange(db, currentChunk, version));
|
||||||
|
|
||||||
|
wait(waitForAll(checkParts));
|
||||||
|
readVersionOut = checkParts.back().get();
|
||||||
|
|
||||||
|
elapsed += timer_monotonic();
|
||||||
|
|
||||||
fmt::print("Blob check complete for [{0} - {1}) @ {2} in {3:.6f} seconds\n",
|
fmt::print("Blob check complete for [{0} - {1}) @ {2} in {3:.6f} seconds\n",
|
||||||
startKey.printable(),
|
startKey.printable(),
|
||||||
endKey.printable(),
|
endKey.printable(),
|
||||||
|
@ -259,6 +259,7 @@ struct BlobManagerStats {
|
|||||||
|
|
||||||
Counter granuleSplits;
|
Counter granuleSplits;
|
||||||
Counter granuleWriteHotSplits;
|
Counter granuleWriteHotSplits;
|
||||||
|
Counter granuleMerges;
|
||||||
Counter ccGranulesChecked;
|
Counter ccGranulesChecked;
|
||||||
Counter ccRowsChecked;
|
Counter ccRowsChecked;
|
||||||
Counter ccBytesChecked;
|
Counter ccBytesChecked;
|
||||||
@ -270,16 +271,27 @@ struct BlobManagerStats {
|
|||||||
Counter granulesPartiallyPurged;
|
Counter granulesPartiallyPurged;
|
||||||
Counter filesPurged;
|
Counter filesPurged;
|
||||||
Future<Void> logger;
|
Future<Void> logger;
|
||||||
|
int64_t activeMerges;
|
||||||
|
|
||||||
// Current stats maintained for a given blob worker process
|
// Current stats maintained for a given blob worker process
|
||||||
explicit BlobManagerStats(UID id, double interval, std::unordered_map<UID, BlobWorkerInterface>* workers)
|
explicit BlobManagerStats(UID id,
|
||||||
|
double interval,
|
||||||
|
int64_t epoch,
|
||||||
|
std::unordered_map<UID, BlobWorkerInterface>* workers,
|
||||||
|
std::unordered_map<Key, bool>* mergeHardBoundaries,
|
||||||
|
std::unordered_map<Key, BlobGranuleMergeBoundary>* mergeBoundaries)
|
||||||
: cc("BlobManagerStats", id.toString()), granuleSplits("GranuleSplits", cc),
|
: cc("BlobManagerStats", id.toString()), granuleSplits("GranuleSplits", cc),
|
||||||
granuleWriteHotSplits("GranuleWriteHotSplits", cc), ccGranulesChecked("CCGranulesChecked", cc),
|
granuleWriteHotSplits("GranuleWriteHotSplits", cc), granuleMerges("GranuleMerges", cc),
|
||||||
ccRowsChecked("CCRowsChecked", cc), ccBytesChecked("CCBytesChecked", cc), ccMismatches("CCMismatches", cc),
|
ccGranulesChecked("CCGranulesChecked", cc), ccRowsChecked("CCRowsChecked", cc),
|
||||||
ccTimeouts("CCTimeouts", cc), ccErrors("CCErrors", cc), purgesProcessed("PurgesProcessed", cc),
|
ccBytesChecked("CCBytesChecked", cc), ccMismatches("CCMismatches", cc), ccTimeouts("CCTimeouts", cc),
|
||||||
|
ccErrors("CCErrors", cc), purgesProcessed("PurgesProcessed", cc),
|
||||||
granulesFullyPurged("GranulesFullyPurged", cc), granulesPartiallyPurged("GranulesPartiallyPurged", cc),
|
granulesFullyPurged("GranulesFullyPurged", cc), granulesPartiallyPurged("GranulesPartiallyPurged", cc),
|
||||||
filesPurged("FilesPurged", cc) {
|
filesPurged("FilesPurged", cc), activeMerges(0) {
|
||||||
specialCounter(cc, "WorkerCount", [workers]() { return workers->size(); });
|
specialCounter(cc, "WorkerCount", [workers]() { return workers->size(); });
|
||||||
|
specialCounter(cc, "Epoch", [epoch]() { return epoch; });
|
||||||
|
specialCounter(cc, "ActiveMerges", [this]() { return this->activeMerges; });
|
||||||
|
specialCounter(cc, "HardBoundaries", [mergeHardBoundaries]() { return mergeHardBoundaries->size(); });
|
||||||
|
specialCounter(cc, "SoftBoundaries", [mergeBoundaries]() { return mergeBoundaries->size(); });
|
||||||
logger = traceCounters("BlobManagerMetrics", id, interval, &cc, "BlobManagerMetrics");
|
logger = traceCounters("BlobManagerMetrics", id, interval, &cc, "BlobManagerMetrics");
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
@ -364,18 +376,23 @@ struct BlobManagerData : NonCopyable, ReferenceCounted<BlobManagerData> {
|
|||||||
Promise<Void> foundBlobWorkers;
|
Promise<Void> foundBlobWorkers;
|
||||||
Promise<Void> doneRecovering;
|
Promise<Void> doneRecovering;
|
||||||
|
|
||||||
int64_t epoch = -1;
|
int64_t epoch;
|
||||||
int64_t seqNo = 1;
|
int64_t seqNo = 1;
|
||||||
|
|
||||||
Promise<Void> iAmReplaced;
|
Promise<Void> iAmReplaced;
|
||||||
|
|
||||||
BlobManagerData(UID id, Reference<AsyncVar<ServerDBInfo> const> dbInfo, Database db, Optional<Key> dcId)
|
BlobManagerData(UID id,
|
||||||
: id(id), db(db), dcId(dcId), stats(id, SERVER_KNOBS->WORKER_LOGGING_INTERVAL, &workersById),
|
Reference<AsyncVar<ServerDBInfo> const> dbInfo,
|
||||||
|
Database db,
|
||||||
|
Optional<Key> dcId,
|
||||||
|
int64_t epoch)
|
||||||
|
: id(id), db(db), dcId(dcId),
|
||||||
|
stats(id, SERVER_KNOBS->WORKER_LOGGING_INTERVAL, epoch, &workersById, &mergeHardBoundaries, &mergeBoundaries),
|
||||||
knownBlobRanges(false, normalKeys.end), tenantData(BGTenantMap(dbInfo)),
|
knownBlobRanges(false, normalKeys.end), tenantData(BGTenantMap(dbInfo)),
|
||||||
mergeCandidates(MergeCandidateInfo(MergeCandidateUnknown), normalKeys.end),
|
mergeCandidates(MergeCandidateInfo(MergeCandidateUnknown), normalKeys.end),
|
||||||
activeGranuleMerges(invalidVersion, normalKeys.end),
|
activeGranuleMerges(invalidVersion, normalKeys.end),
|
||||||
concurrentMergeChecks(SERVER_KNOBS->BLOB_MANAGER_CONCURRENT_MERGE_CHECKS),
|
concurrentMergeChecks(SERVER_KNOBS->BLOB_MANAGER_CONCURRENT_MERGE_CHECKS),
|
||||||
restartRecruiting(SERVER_KNOBS->DEBOUNCE_RECRUITING_DELAY), recruitingStream(0) {}
|
restartRecruiting(SERVER_KNOBS->DEBOUNCE_RECRUITING_DELAY), recruitingStream(0), epoch(epoch) {}
|
||||||
|
|
||||||
// only initialize blob store if actually needed
|
// only initialize blob store if actually needed
|
||||||
void initBStore() {
|
void initBStore() {
|
||||||
@ -1874,6 +1891,7 @@ ACTOR Future<Void> finishMergeGranules(Reference<BlobManagerData> bmData,
|
|||||||
std::vector<UID> parentGranuleIDs,
|
std::vector<UID> parentGranuleIDs,
|
||||||
std::vector<Key> parentGranuleRanges,
|
std::vector<Key> parentGranuleRanges,
|
||||||
std::vector<Version> parentGranuleStartVersions) {
|
std::vector<Version> parentGranuleStartVersions) {
|
||||||
|
++bmData->stats.activeMerges;
|
||||||
|
|
||||||
// wait for BM to be fully recovered before starting actual merges
|
// wait for BM to be fully recovered before starting actual merges
|
||||||
wait(bmData->doneRecovering.getFuture());
|
wait(bmData->doneRecovering.getFuture());
|
||||||
@ -1920,6 +1938,8 @@ ACTOR Future<Void> finishMergeGranules(Reference<BlobManagerData> bmData,
|
|||||||
BoundaryEvaluation(bmData->epoch, seqnoForEval, BoundaryEvalType::MERGE, 0, 0));
|
BoundaryEvaluation(bmData->epoch, seqnoForEval, BoundaryEvalType::MERGE, 0, 0));
|
||||||
bmData->setMergeCandidate(mergeRange, MergeCandidateMerging);
|
bmData->setMergeCandidate(mergeRange, MergeCandidateMerging);
|
||||||
|
|
||||||
|
--bmData->stats.activeMerges;
|
||||||
|
|
||||||
return Void();
|
return Void();
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1937,6 +1957,8 @@ ACTOR Future<Void> doMerge(Reference<BlobManagerData> bmData,
|
|||||||
}
|
}
|
||||||
ranges.push_back(std::get<1>(toMerge.back()).end);
|
ranges.push_back(std::get<1>(toMerge.back()).end);
|
||||||
|
|
||||||
|
++bmData->stats.granuleMerges;
|
||||||
|
|
||||||
try {
|
try {
|
||||||
std::pair<UID, Version> persistMerge =
|
std::pair<UID, Version> persistMerge =
|
||||||
wait(persistMergeGranulesStart(bmData, mergeRange, ids, ranges, startVersions));
|
wait(persistMergeGranulesStart(bmData, mergeRange, ids, ranges, startVersions));
|
||||||
@ -4251,7 +4273,8 @@ ACTOR Future<Void> blobManager(BlobManagerInterface bmInterf,
|
|||||||
makeReference<BlobManagerData>(deterministicRandom()->randomUniqueID(),
|
makeReference<BlobManagerData>(deterministicRandom()->randomUniqueID(),
|
||||||
dbInfo,
|
dbInfo,
|
||||||
openDBOnServer(dbInfo, TaskPriority::DefaultEndpoint, LockAware::True),
|
openDBOnServer(dbInfo, TaskPriority::DefaultEndpoint, LockAware::True),
|
||||||
bmInterf.locality.dcId());
|
bmInterf.locality.dcId(),
|
||||||
|
epoch);
|
||||||
|
|
||||||
state Future<Void> collection = actorCollection(self->addActor.getFuture());
|
state Future<Void> collection = actorCollection(self->addActor.getFuture());
|
||||||
|
|
||||||
@ -4260,8 +4283,6 @@ ACTOR Future<Void> blobManager(BlobManagerInterface bmInterf,
|
|||||||
}
|
}
|
||||||
TraceEvent("BlobManagerInit", bmInterf.id()).detail("Epoch", epoch).log();
|
TraceEvent("BlobManagerInit", bmInterf.id()).detail("Epoch", epoch).log();
|
||||||
|
|
||||||
self->epoch = epoch;
|
|
||||||
|
|
||||||
// although we start the recruiter, we wait until existing workers are ack'd
|
// although we start the recruiter, we wait until existing workers are ack'd
|
||||||
auto recruitBlobWorker = IAsyncListener<RequestStream<RecruitBlobWorkerRequest>>::create(
|
auto recruitBlobWorker = IAsyncListener<RequestStream<RecruitBlobWorkerRequest>>::create(
|
||||||
dbInfo, [](auto const& info) { return info.clusterInterface.recruitBlobWorker; });
|
dbInfo, [](auto const& info) { return info.clusterInterface.recruitBlobWorker; });
|
||||||
|
Loading…
x
Reference in New Issue
Block a user