/* * Copyright (c) Meta Platforms, Inc. and affiliates. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include #include #include namespace apache::thrift { namespace detail { THRIFT_PLUGGABLE_FUNC_REGISTER( folly::observer::Observer, makeCPUConcurrencyControllerConfig, ThriftServer*) { return folly::observer::makeStaticObserver( CPUConcurrencyController::Config{}); } THRIFT_PLUGGABLE_FUNC_REGISTER( int64_t, getCPULoadCounter, std::chrono::milliseconds, CPULoadSource) { return -1; } } // namespace detail CPUConcurrencyController::CPUConcurrencyController( folly::observer::Observer config, apache::thrift::server::ServerConfigs& serverConfigs, apache::thrift::ThriftServerConfig& thriftServerConfig, std::optional loadFunc) : config_{(*config).getShared()}, enabled_{(*config_.rlock())->enabled()}, method_{(*config_.rlock())->method}, serverConfigs_(serverConfigs), thriftServerConfig_(thriftServerConfig), loadFunc_(std::move(loadFunc)) { scheduler_.setThreadName("CPUConcurrencyController-loop"); scheduler_.start(); configSchedulerCallback_ = config.addCallback( [this](folly::observer::Snapshot configSnapshot) { auto newConfig = configSnapshot.getShared(); this->config_.withWLock([&newConfig, this](auto& config) { config = newConfig; this->enabled_.store(config->enabled(), std::memory_order_relaxed); this->method_.store(config->method, std::memory_order_relaxed); }); this->eventHandler_.withRLock([&newConfig](const auto& eventHandler) { if (eventHandler) { eventHandler->configUpdated(newConfig); } }); this->cancel(); this->schedule(std::move(newConfig)); }); } CPUConcurrencyController::~CPUConcurrencyController() { // Cancel to avoid using CPUConcurrencyController members while its // partially destructed configSchedulerCallback_.cancel(); cancel(); } void CPUConcurrencyController::setEventHandler( std::shared_ptr eventHandler) { eventHandler_.withWLock( [newEventHandler = std::move(eventHandler), currentConfig = config()](auto& eventHandler) mutable { eventHandler = std::move(newEventHandler); if (eventHandler) { eventHandler->configUpdated(std::move(currentConfig)); } }); } void CPUConcurrencyController::requestStarted() { if (!enabled_fast()) { return; } totalRequestCount_ += 1; } bool CPUConcurrencyController::requestShed(std::optional method) { if (!enabled_fast()) { return false; } if (method && *method != method_.load(std::memory_order_relaxed)) { return false; } recentShedRequest_.store(true); return true; } void CPUConcurrencyController::cycleOnce() { // Get a snapshot of the current config auto config = this->config(); auto eventHandler = eventHandler_.copy(); if (!config->enabled()) { return; } auto limit = this->getLimit(config); auto currentLimitUsage = this->getLimitUsage(config); auto load = getLoadInternal(config); if (eventHandler) { eventHandler->onCycle(limit, currentLimitUsage, load); } if (load >= config->cpuTarget) { lastOverloadStart_ = std::chrono::steady_clock::now(); auto newLim = limit - std::max( static_cast(limit * config->decreaseMultiplier), 1); this->setLimit( config, std::max(newLim, config->concurrencyLowerBound)); if (eventHandler) { eventHandler->limitDecreased(); } } else { if (currentLimitUsage == 0 || load <= 0) { return; } // Estimate stable concurrency only if we haven't been overloaded recently // (and thus current concurrency/CPU is not a good indicator). if (!isRefractoryPeriodInternal(config) && config->collectionSampleSize > stableConcurrencySamples_.size()) { auto concurrencyEstimate = static_cast(currentLimitUsage) / load * config->cpuTarget; stableConcurrencySamples_.push_back( config->initialEstimateFactor * concurrencyEstimate); if (stableConcurrencySamples_.size() >= config->collectionSampleSize) { // Take percentile to hopefully account for inaccuracies. We don't // need a very accurate value as we will adjust this later in the // algorithm. // // The purpose of stable concurrency estimate is to optimistically // avoid causing too much request ingestion during initial overload, and // thus causing high tail latencies during initial overload. auto pct = stableConcurrencySamples_.begin() + static_cast( stableConcurrencySamples_.size() * config->initialEstimatePercentile); std::nth_element( stableConcurrencySamples_.begin(), pct, stableConcurrencySamples_.end()); auto result = std::clamp( *pct, config->concurrencyLowerBound, getConcurrencyUpperBoundInternal(config)); stableEstimate_.store(result, std::memory_order_relaxed); this->setLimit(config, result); if (eventHandler) { if (result > limit) { eventHandler->limitIncreased(); } else if (result < limit) { eventHandler->limitDecreased(); } } return; } } // We prevent unbounded increase of limits by only changing when // necessary (i.e., we are getting near breaching existing limit). bool nearExistingLimit = !config->bumpOnError && (currentLimitUsage >= (1.0 - config->increaseDistanceRatio) * limit); // Bump concurrency limit if we saw some load shedding errors recently. bool recentLoadShed = config->bumpOnError && recentShedRequest_.exchange(false); // We try and push the limit back to stable estimate if we are not // overloaded as after an overload event the limit will be set aggressively // and may cause steady-state load shedding due to bursty traffic. bool shouldConvergeStable = !isRefractoryPeriodInternal(config) && limit < getStableEstimate(); if (nearExistingLimit || recentLoadShed || shouldConvergeStable) { auto newLim = limit + std::max( static_cast(limit * config->additiveMultiplier), 1); this->setLimit( config, std::min(getConcurrencyUpperBoundInternal(config), newLim)); if (eventHandler) { eventHandler->limitIncreased(); } } } } void CPUConcurrencyController::schedule(std::shared_ptr config) { using time_point = std::chrono::steady_clock::time_point; if (!config->enabled()) { return; } LOG(INFO) << "Enabling CPUConcurrencyController: " << config->describe(); thriftServerConfig_.setMaxRequests( activeRequestsLimit_.getObserver(), apache::thrift::AttributeSource::OVERRIDE_INTERNAL); thriftServerConfig_.setMaxQps( qpsLimit_.getObserver(), apache::thrift::AttributeSource::OVERRIDE_INTERNAL); this->setLimit(config, getConcurrencyUpperBoundInternal(config)); scheduler_.addFunctionGenericNextRunTimeFunctor( [this] { this->cycleOnce(); }, [config](time_point, time_point now) { return now + config->refreshPeriodMs; }, "cpu-shed-loop", "cpu-shed-interval", config->refreshPeriodMs); } void CPUConcurrencyController::cancel() { scheduler_.cancelAllFunctionsAndWait(); activeRequestsLimit_.setValue(std::nullopt); qpsLimit_.setValue(std::nullopt); dryRunLimit_ = 0; stableConcurrencySamples_.clear(); stableEstimate_.exchange(-1); } bool CPUConcurrencyController::enabled_fast() const { return enabled_.load(std::memory_order_relaxed); } uint32_t CPUConcurrencyController::getLimit( const std::shared_ptr& config) const { uint32_t limit = 0; if (config->mode == Mode::DRY_RUN) { limit = dryRunLimit_; } else { switch (config->method) { case Method::MAX_REQUESTS: // Using ServiceConfigs instead of ThriftServerConfig, because the value // may come from AdaptiveConcurrencyController limit = serverConfigs_.getMaxRequests(); break; case Method::MAX_QPS: limit = thriftServerConfig_.getMaxQps().get(); break; default: DCHECK(false); } } // Fallback to concurrency upper bound if no limit is set yet. // This is most sensible value until we collect enough samples // to estimate a better upper bound; return limit ? limit : getConcurrencyUpperBoundInternal(config); } void CPUConcurrencyController::setLimit( const std::shared_ptr& config, uint32_t newLimit) { if (config->mode == Mode::DRY_RUN) { dryRunLimit_ = newLimit; } else if (config->mode == Mode::ENABLED) { switch (config->method) { case Method::MAX_REQUESTS: activeRequestsLimit_.setValue(newLimit); break; case Method::MAX_QPS: qpsLimit_.setValue(newLimit); break; default: DCHECK(false); } } } uint32_t CPUConcurrencyController::getLimitUsage( const std::shared_ptr& config) { using namespace std::chrono; switch (config->method) { case Method::MAX_REQUESTS: // Note: estimating concurrency from this is fairly lossy as it's a // gauge metric and we can't use techniques to measure it over a duration. // We may be able to get much better estimates if we switch to use QPS // and a token bucket for rate limiting. // TODO: We should exclude fb303 methods. return serverConfigs_.getActiveRequests(); case Method::MAX_QPS: { auto now = steady_clock::now(); auto milliSince = duration_cast(now - lastTotalRequestReset_).count(); if (milliSince == 0) { return 0; } auto totalRequestSince = totalRequestCount_.load(); totalRequestCount_ = 0; lastTotalRequestReset_ = now; return totalRequestSince * 1000L / milliSince; } default: DCHECK(false); return 0; } } bool CPUConcurrencyController::isRefractoryPeriodInternal( const std::shared_ptr& config) const { return (std::chrono::steady_clock::now() - lastOverloadStart_) <= std::chrono::milliseconds(config->refractoryPeriodMs); } int64_t CPUConcurrencyController::getLoadInternal( const std::shared_ptr& config) const { auto& refreshPeriodMs = config->refreshPeriodMs; auto& cpuLoadSource = config->cpuLoadSource; if (loadFunc_) { return std::clamp( (*loadFunc_)(refreshPeriodMs, cpuLoadSource), 0, 100); } return std::clamp( detail::getCPULoadCounter(refreshPeriodMs, cpuLoadSource), 0, 100); } uint32_t CPUConcurrencyController::getConcurrencyUpperBoundInternal( const std::shared_ptr& config) const { if (std::holds_alternative( config->concurrencyUpperBound)) { // Use static limit as concurrencyUpperBound return config->method == Method::MAX_REQUESTS ? thriftServerConfig_.getMaxRequests().get() : thriftServerConfig_.getMaxQps().get(); } return std::max(std::get(config->concurrencyUpperBound), 0); } /** * CPUConcurrencyController Config helper methods */ bool CPUConcurrencyController::Config::enabled() const { return mode != Mode::DISABLED; } std::string_view CPUConcurrencyController::Config::modeName() const { switch (mode) { case Mode::DISABLED: return "DISABLED"; case Mode::DRY_RUN: return "DRY_RUN"; case Mode::ENABLED: return "ENABLED"; } folly::assume_unreachable(); } std::string_view CPUConcurrencyController::Config::methodName() const { switch (method) { // TODO(sazonovk): What's the effect of changing these strings? case Method::MAX_REQUESTS: return "MAX_REQUESTS"; case Method::MAX_QPS: return "MAX_QPS"; } folly::assume_unreachable(); } std::string_view CPUConcurrencyController::Config::cpuLoadSourceName() const { switch (cpuLoadSource) { case CPULoadSource::CONTAINER_AND_HOST: return "CONTAINER_AND_HOST"; case CPULoadSource::CONTAINER_ONLY: return "CONTAINER_ONLY"; case CPULoadSource::HOST_ONLY: return "HOST_ONLY"; } folly::assume_unreachable(); } std::string_view CPUConcurrencyController::Config::concurrencyUnit() const { switch (method) { // TODO(sazonovk): What's the effect of changing these strings? case Method::MAX_REQUESTS: return "maxRequests"; case Method::MAX_QPS: return "maxQps"; } folly::assume_unreachable(); } std::string CPUConcurrencyController::Config::concurrencyUpperBoundName() const { return folly::variant_match( concurrencyUpperBound, [](const UseStaticLimit&) { return std::string{"UseStaticLimit"}; }, [](const int32_t& i) { return fmt::format("{}", i); }); } std::string CPUConcurrencyController::Config::describe() const { return fmt::format( "CPUConcurrencyController Config: " "Mode: {}, " "Method: {}, " "CPU Target: {}, " "Refresh Period (ms): {}, " "Concurrency Upper Bound({}): {}", modeName(), methodName(), cpuTarget, refreshPeriodMs.count(), concurrencyUnit(), concurrencyUpperBoundName()); } serverdbginfo::CPUConcurrencyControllerDbgInfo CPUConcurrencyController::getDbgInfo() const { serverdbginfo::CPUConcurrencyControllerDbgInfo info; auto configLocal = config(); info.mode() = configLocal->modeName(); info.method() = configLocal->methodName(); info.cpuTarget() = configLocal->cpuTarget; info.cpuLoadSource() = configLocal->cpuLoadSourceName(); info.refreshPeriodMs() = std::chrono::duration_cast( configLocal->refreshPeriodMs) .count(); info.additiveMultiplier() = configLocal->additiveMultiplier; info.decreaseMultiplier() = configLocal->decreaseMultiplier; info.increaseDistanceRatio() = configLocal->increaseDistanceRatio; info.bumpOnError() = configLocal->bumpOnError; info.refractoryPeriodMs() = std::chrono::duration_cast( configLocal->refractoryPeriodMs) .count(); info.initialEstimateFactor() = configLocal->initialEstimateFactor; info.initialEstimatePercentile() = configLocal->initialEstimatePercentile; info.collectionSampleSize() = configLocal->collectionSampleSize; info.concurrencyUpperBound() = folly::variant_match( configLocal->concurrencyUpperBound, [](const CPUConcurrencyController::Config::UseStaticLimit&) { return -1; }, [](const int32_t& i) { return i; }); info.concurrencyLowerBound() = configLocal->concurrencyLowerBound; info.cpuLoad() = getLoadInternal(configLocal); return info; } } // namespace apache::thrift