mirror of
https://github.com/apple/foundationdb.git
synced 2025-06-02 11:15:50 +08:00
Merge pull request #6082 from sfc-gh-tclinkenbeard/fix-unused-warnings
Fix `-Wunused-but-set-variable` warnings
This commit is contained in:
commit
ec02d72338
@ -2998,15 +2998,15 @@ static std::vector<std::vector<StringRef>> parseLine(std::string& line, bool& er
|
||||
|
||||
static void addKeyRange(std::string optionValue, Standalone<VectorRef<KeyRangeRef>>& keyRanges) {
|
||||
bool err = false, partial = false;
|
||||
int tokenArray = 0, tokenIndex = 0;
|
||||
int tokenArray = 0;
|
||||
|
||||
auto parsed = parseLine(optionValue, err, partial);
|
||||
|
||||
for (auto tokens : parsed) {
|
||||
tokenArray++;
|
||||
tokenIndex = 0;
|
||||
|
||||
/*
|
||||
int tokenIndex = 0;
|
||||
for (auto token : tokens)
|
||||
{
|
||||
tokenIndex++;
|
||||
|
@ -454,7 +454,7 @@ ConfigureAutoResult parseConfig(StatusObject const& status) {
|
||||
|
||||
result.auto_replication = result.old_replication;
|
||||
|
||||
int storage_replication;
|
||||
[[maybe_unused]] int storage_replication;
|
||||
int log_replication;
|
||||
if (result.old_replication == "single") {
|
||||
result.auto_replication = "double";
|
||||
|
@ -1151,7 +1151,7 @@ void testPathOps() {
|
||||
std::string cwd = abspath(".", true);
|
||||
|
||||
// Create some symlinks and test resolution (or non-resolution) of them
|
||||
int rc;
|
||||
[[maybe_unused]] int rc;
|
||||
// Ignoring return codes, if symlinks fail tests below will fail
|
||||
rc = unlink("simfdb/backups/four");
|
||||
rc = unlink("simfdb/backups/five");
|
||||
|
@ -1951,7 +1951,7 @@ ACTOR Future<Void> handleBlobGranuleFileRequest(Reference<BlobWorkerData> bwData
|
||||
j--;
|
||||
}
|
||||
j++;
|
||||
Version latestDeltaVersion = invalidVersion;
|
||||
[[maybe_unused]] Version latestDeltaVersion = invalidVersion;
|
||||
while (j <= i) {
|
||||
BlobFileIndex deltaF = chunkFiles.deltaFiles[j];
|
||||
chunk.deltaFiles.emplace_back_deep(rep.arena, deltaF.filename, deltaF.offset, deltaF.length);
|
||||
|
@ -2195,7 +2195,6 @@ struct DDTeamCollection : ReferenceCounted<DDTeamCollection> {
|
||||
ASSERT_WE_THINK(machine_info.size() > 0 || server_info.size() == 0);
|
||||
ASSERT_WE_THINK(SERVER_KNOBS->DESIRED_TEAMS_PER_SERVER >= 1 && configuration.storageTeamSize >= 1);
|
||||
|
||||
int addedMachineTeams = 0;
|
||||
int addedTeams = 0;
|
||||
|
||||
// Exclude machine teams who have members in the wrong configuration.
|
||||
@ -2210,15 +2209,18 @@ struct DDTeamCollection : ReferenceCounted<DDTeamCollection> {
|
||||
int machineTeamsToBuild = std::max(
|
||||
0, std::min(desiredMachineTeams - healthyMachineTeamCount, maxMachineTeams - totalMachineTeamCount));
|
||||
|
||||
TraceEvent("BuildMachineTeams")
|
||||
.detail("TotalHealthyMachine", totalHealthyMachineCount)
|
||||
.detail("HealthyMachineTeamCount", healthyMachineTeamCount)
|
||||
.detail("DesiredMachineTeams", desiredMachineTeams)
|
||||
.detail("MaxMachineTeams", maxMachineTeams)
|
||||
.detail("MachineTeamsToBuild", machineTeamsToBuild);
|
||||
// Pre-build all machine teams until we have the desired number of machine teams
|
||||
if (machineTeamsToBuild > 0 || notEnoughMachineTeamsForAMachine()) {
|
||||
addedMachineTeams = addBestMachineTeams(machineTeamsToBuild);
|
||||
{
|
||||
TraceEvent te("BuildMachineTeams");
|
||||
te.detail("TotalHealthyMachine", totalHealthyMachineCount)
|
||||
.detail("HealthyMachineTeamCount", healthyMachineTeamCount)
|
||||
.detail("DesiredMachineTeams", desiredMachineTeams)
|
||||
.detail("MaxMachineTeams", maxMachineTeams)
|
||||
.detail("MachineTeamsToBuild", machineTeamsToBuild);
|
||||
// Pre-build all machine teams until we have the desired number of machine teams
|
||||
if (machineTeamsToBuild > 0 || notEnoughMachineTeamsForAMachine()) {
|
||||
auto addedMachineTeams = addBestMachineTeams(machineTeamsToBuild);
|
||||
te.detail("MachineTeamsAdded", addedMachineTeams);
|
||||
}
|
||||
}
|
||||
|
||||
while (addedTeams < teamsToBuild || notEnoughTeamsForAServer()) {
|
||||
|
@ -180,13 +180,10 @@ ACTOR Future<Void> resolveBatch(Reference<Resolver> self, ResolveTransactionBatc
|
||||
// Detect conflicts
|
||||
double expire = now() + SERVER_KNOBS->SAMPLE_EXPIRATION_TIME;
|
||||
ConflictBatch conflictBatch(self->conflictSet, &reply.conflictingKeyRangeMap, &reply.arena);
|
||||
int keys = 0;
|
||||
for (int t = 0; t < req.transactions.size(); t++) {
|
||||
conflictBatch.addTransaction(req.transactions[t]);
|
||||
self->resolvedReadConflictRanges += req.transactions[t].read_conflict_ranges.size();
|
||||
self->resolvedWriteConflictRanges += req.transactions[t].write_conflict_ranges.size();
|
||||
keys += req.transactions[t].write_conflict_ranges.size() * 2 +
|
||||
req.transactions[t].read_conflict_ranges.size() * 2;
|
||||
|
||||
if (self->resolverCount > 1) {
|
||||
for (auto it : req.transactions[t].write_conflict_ranges)
|
||||
|
@ -166,7 +166,7 @@ void art_tree::art_bound_iterative(art_node* n, const KeyRef& k, int depth, art_
|
||||
|
||||
static stack_entry arena[ART_MAX_KEY_LEN]; // Single threaded implementation.
|
||||
|
||||
stack_entry *head = nullptr, *tmp, *curr_arena = arena;
|
||||
stack_entry *head = nullptr, *curr_arena = arena;
|
||||
int ret;
|
||||
art_node** child;
|
||||
unsigned char* key = (unsigned char*)k.begin();
|
||||
@ -198,7 +198,6 @@ void art_tree::art_bound_iterative(art_node* n, const KeyRef& k, int depth, art_
|
||||
*result = minimum(next);
|
||||
break;
|
||||
} else {
|
||||
tmp = head;
|
||||
head = head->prev;
|
||||
}
|
||||
}
|
||||
|
@ -1598,7 +1598,8 @@ struct ConsistencyCheckWorkload : TestWorkload {
|
||||
for (int j = 0; j < storageServers.size(); j++)
|
||||
storageServerSizes[storageServers[j]] += shardBytes;
|
||||
|
||||
bool hasValidEstimate = estimatedBytes.size() > 0;
|
||||
// FIXME: Where is this intended to be used?
|
||||
[[maybe_unused]] bool hasValidEstimate = estimatedBytes.size() > 0;
|
||||
|
||||
// If the storage servers' sampled estimate of shard size is different from ours
|
||||
if (self->performQuiescentChecks) {
|
||||
|
@ -21,10 +21,10 @@ void getVersionAndnumTags(TraceEventFields md, Version& version, int& numTags) {
|
||||
}
|
||||
|
||||
void getTagAndDurableVersion(TraceEventFields md, Version version, Tag& tag, Version& durableVersion) {
|
||||
Version verifyVersion;
|
||||
durableVersion = -1;
|
||||
|
||||
verifyVersion = boost::lexical_cast<int64_t>(md.getValue("Version"));
|
||||
// verify version:
|
||||
boost::lexical_cast<int64_t>(md.getValue("Version"));
|
||||
std::string tagString = md.getValue("Tag");
|
||||
int colon = tagString.find_first_of(':');
|
||||
std::string localityString = tagString.substr(0, colon);
|
||||
@ -40,11 +40,11 @@ void getMinAndMaxTLogVersions(TraceEventFields md,
|
||||
Tag tag,
|
||||
Version& minTLogVersion,
|
||||
Version& maxTLogVersion) {
|
||||
Version verifyVersion;
|
||||
Tag verifyTag;
|
||||
minTLogVersion = maxTLogVersion = -1;
|
||||
|
||||
verifyVersion = boost::lexical_cast<int64_t>(md.getValue("Version"));
|
||||
// verify version:
|
||||
boost::lexical_cast<int64_t>(md.getValue("Version"));
|
||||
std::string tagString = md.getValue("Tag");
|
||||
int colon = tagString.find_first_of(':');
|
||||
std::string localityString = tagString.substr(0, colon);
|
||||
|
@ -1509,7 +1509,7 @@ void Net2::run() {
|
||||
taskBegin = timer_monotonic();
|
||||
numYields = 0;
|
||||
TaskPriority minTaskID = TaskPriority::Max;
|
||||
int queueSize = ready.size();
|
||||
[[maybe_unused]] int queueSize = ready.size();
|
||||
|
||||
FDB_TRACE_PROBE(run_loop_tasks_start, queueSize);
|
||||
while (!ready.empty()) {
|
||||
|
Loading…
x
Reference in New Issue
Block a user