blob: 0a5c43a99be5fccc8cfc646dffd29b8a22828d68 [file] [log] [blame]
/*
* Copyright 2021 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#undef LOG_TAG
#define LOG_TAG "Planner"
// #define LOG_NDEBUG 0
#define ATRACE_TAG ATRACE_TAG_GRAPHICS
#include <android-base/properties.h>
#include <common/FlagManager.h>
#include <compositionengine/impl/planner/Flattener.h>
#include <compositionengine/impl/planner/LayerState.h>
#include <gui/TraceUtils.h>
using time_point = std::chrono::steady_clock::time_point;
using namespace std::chrono_literals;
namespace android::compositionengine::impl::planner {
namespace {
// True if the underlying layer stack is the same modulo state that would be expected to be
// different like specific buffers, false otherwise.
bool isSameStack(const std::vector<const LayerState*>& incomingLayers,
const std::vector<CachedSet>& cachedSets) {
std::vector<const LayerState*> existingLayers;
for (auto& cachedSet : cachedSets) {
for (auto& layer : cachedSet.getConstituentLayers()) {
existingLayers.push_back(layer.getState());
}
}
if (incomingLayers.size() != existingLayers.size()) {
return false;
}
for (size_t i = 0; i < incomingLayers.size(); i++) {
// Checking the IDs here is very strict, but we do this as otherwise we may mistakenly try
// to access destroyed OutputLayers later on.
if (incomingLayers[i]->getId() != existingLayers[i]->getId()) {
return false;
}
// Do not unflatten if source crop is only moved.
if (FlagManager::getInstance().cache_when_source_crop_layer_only_moved() &&
incomingLayers[i]->isSourceCropSizeEqual(*(existingLayers[i])) &&
incomingLayers[i]->getDifferingFields(*(existingLayers[i])) ==
LayerStateField::SourceCrop) {
continue;
}
if (incomingLayers[i]->getDifferingFields(*(existingLayers[i])) != LayerStateField::None) {
return false;
}
}
return true;
}
} // namespace
Flattener::Flattener(renderengine::RenderEngine& renderEngine, const Tunables& tunables)
: mRenderEngine(renderEngine), mTunables(tunables), mTexturePool(mRenderEngine) {}
NonBufferHash Flattener::flattenLayers(const std::vector<const LayerState*>& layers,
NonBufferHash hash, time_point now) {
ATRACE_CALL();
const size_t unflattenedDisplayCost = calculateDisplayCost(layers);
mUnflattenedDisplayCost += unflattenedDisplayCost;
// We invalidate the layer cache if:
// 1. We're not tracking any layers, or
// 2. The last seen hashed geometry changed between frames, or
// 3. A stricter equality check demonstrates that the layer stack really did change, since the
// hashed geometry does not guarantee uniqueness.
if (mCurrentGeometry != hash || (!mLayers.empty() && !isSameStack(layers, mLayers))) {
resetActivities(hash, now);
mFlattenedDisplayCost += unflattenedDisplayCost;
return hash;
}
++mInitialLayerCounts[layers.size()];
// Only buildCachedSets if these layers are already stored in mLayers.
// Otherwise (i.e. mergeWithCachedSets returns false), the time has not
// changed, so buildCachedSets will never find any runs.
const bool alreadyHadCachedSets = mergeWithCachedSets(layers, now);
++mFinalLayerCounts[mLayers.size()];
if (alreadyHadCachedSets) {
buildCachedSets(now);
hash = computeLayersHash();
}
return hash;
}
void Flattener::renderCachedSets(
const OutputCompositionState& outputState,
std::optional<std::chrono::steady_clock::time_point> renderDeadline,
bool deviceHandlesColorTransform) {
ATRACE_CALL();
if (!mNewCachedSet) {
return;
}
// Ensure that a cached set has a valid buffer first
if (mNewCachedSet->hasRenderedBuffer()) {
ATRACE_NAME("mNewCachedSet->hasRenderedBuffer()");
return;
}
const auto now = std::chrono::steady_clock::now();
// If we have a render deadline, and the flattener is configured to skip rendering if we don't
// have enough time, then we skip rendering the cached set if we think that we'll steal too much
// time from the next frame.
if (renderDeadline && mTunables.mRenderScheduling) {
if (const auto estimatedRenderFinish =
now + mTunables.mRenderScheduling->cachedSetRenderDuration;
estimatedRenderFinish > *renderDeadline) {
mNewCachedSet->incrementSkipCount();
if (mNewCachedSet->getSkipCount() <=
mTunables.mRenderScheduling->maxDeferRenderAttempts) {
ATRACE_FORMAT("DeadlinePassed: exceeded deadline by: %d us",
std::chrono::duration_cast<std::chrono::microseconds>(
estimatedRenderFinish - *renderDeadline)
.count());
return;
} else {
ATRACE_NAME("DeadlinePassed: exceeded max skips");
}
}
}
mNewCachedSet->render(mRenderEngine, mTexturePool, outputState, deviceHandlesColorTransform);
}
void Flattener::dumpLayers(std::string& result) const {
result.append(" Current layers:");
for (const CachedSet& layer : mLayers) {
result.append("\n");
layer.dump(result);
}
}
void Flattener::dump(std::string& result) const {
const auto now = std::chrono::steady_clock::now();
base::StringAppendF(&result, "Flattener state:\n");
result.append("\n Statistics:\n");
result.append(" Display cost (in screen-size buffers):\n");
const size_t displayArea = static_cast<size_t>(mDisplaySize.width * mDisplaySize.height);
base::StringAppendF(&result, " Unflattened: %.2f\n",
static_cast<float>(mUnflattenedDisplayCost) / displayArea);
base::StringAppendF(&result, " Flattened: %.2f\n",
static_cast<float>(mFlattenedDisplayCost) / displayArea);
const auto compareLayerCounts = [](const std::pair<size_t, size_t>& left,
const std::pair<size_t, size_t>& right) {
return left.first < right.first;
};
const size_t maxLayerCount = mInitialLayerCounts.empty()
? 0u
: std::max_element(mInitialLayerCounts.cbegin(), mInitialLayerCounts.cend(),
compareLayerCounts)
->first;
result.append("\n Initial counts:\n");
for (size_t count = 1; count < maxLayerCount; ++count) {
size_t initial = mInitialLayerCounts.count(count) > 0 ? mInitialLayerCounts.at(count) : 0;
base::StringAppendF(&result, " % 2zd: %zd\n", count, initial);
}
result.append("\n Final counts:\n");
for (size_t count = 1; count < maxLayerCount; ++count) {
size_t final = mFinalLayerCounts.count(count) > 0 ? mFinalLayerCounts.at(count) : 0;
base::StringAppendF(&result, " % 2zd: %zd\n", count, final);
}
base::StringAppendF(&result, "\n Cached sets created: %zd\n", mCachedSetCreationCount);
base::StringAppendF(&result, " Cost: %.2f\n",
static_cast<float>(mCachedSetCreationCost) / displayArea);
const auto lastUpdate =
std::chrono::duration_cast<std::chrono::milliseconds>(now - mLastGeometryUpdate);
base::StringAppendF(&result, "\n Current hash %016zx, last update %sago\n\n", mCurrentGeometry,
durationString(lastUpdate).c_str());
dumpLayers(result);
base::StringAppendF(&result, "\n");
mTexturePool.dump(result);
}
size_t Flattener::calculateDisplayCost(const std::vector<const LayerState*>& layers) const {
Region coveredRegion;
size_t displayCost = 0;
bool hasClientComposition = false;
for (const LayerState* layer : layers) {
coveredRegion.orSelf(layer->getDisplayFrame());
// Regardless of composition type, we always have to read each input once
displayCost += static_cast<size_t>(layer->getDisplayFrame().width() *
layer->getDisplayFrame().height());
hasClientComposition |= layer->getCompositionType() ==
aidl::android::hardware::graphics::composer3::Composition::CLIENT;
}
if (hasClientComposition) {
// If there is client composition, the client target buffer has to be both written by the
// GPU and read by the DPU, so we pay its cost twice
displayCost += 2 *
static_cast<size_t>(coveredRegion.bounds().width() *
coveredRegion.bounds().height());
}
return displayCost;
}
void Flattener::resetActivities(NonBufferHash hash, time_point now) {
ALOGV("[%s]", __func__);
mCurrentGeometry = hash;
mLastGeometryUpdate = now;
for (const CachedSet& cachedSet : mLayers) {
if (cachedSet.getLayerCount() > 1) {
++mInvalidatedCachedSetAges[cachedSet.getAge()];
}
}
mLayers.clear();
if (mNewCachedSet) {
++mInvalidatedCachedSetAges[mNewCachedSet->getAge()];
mNewCachedSet = std::nullopt;
}
}
NonBufferHash Flattener::computeLayersHash() const{
size_t hash = 0;
for (const auto& layer : mLayers) {
android::hashCombineSingleHashed(hash, layer.getNonBufferHash());
}
return hash;
}
// Only called if the geometry matches the last frame. Return true if mLayers
// was already populated with these layers, i.e. on the second and following
// calls with the same geometry.
bool Flattener::mergeWithCachedSets(const std::vector<const LayerState*>& layers, time_point now) {
ATRACE_CALL();
std::vector<CachedSet> merged;
if (mLayers.empty()) {
merged.reserve(layers.size());
for (const LayerState* layer : layers) {
merged.emplace_back(layer, now);
mFlattenedDisplayCost += merged.back().getDisplayCost();
}
mLayers = std::move(merged);
return false;
}
// the compiler should strip out the following no-op loops when ALOGV is off
ALOGV("[%s] Incoming layers:", __func__);
for (const LayerState* layer : layers) {
ALOGV("%s", layer->getName().c_str());
}
ALOGV("[%s] Current layers:", __func__);
for (const CachedSet& layer : mLayers) {
const auto dumper = [&] {
std::string dump;
layer.dump(dump);
return dump;
};
ALOGV("%s", dumper().c_str());
}
auto currentLayerIter = mLayers.begin();
auto incomingLayerIter = layers.begin();
// If not null, this represents the layer that is blurring the layer before
// currentLayerIter. The blurring was stored in the override buffer, so the
// layer that requests the blur no longer needs to do any blurring.
compositionengine::OutputLayer* priorBlurLayer = nullptr;
while (incomingLayerIter != layers.end()) {
if (mNewCachedSet &&
mNewCachedSet->getFirstLayer().getState()->getId() == (*incomingLayerIter)->getId()) {
if (mNewCachedSet->hasBufferUpdate()) {
ALOGV("[%s] Dropping new cached set", __func__);
++mInvalidatedCachedSetAges[0];
mNewCachedSet = std::nullopt;
} else if (mNewCachedSet->hasReadyBuffer()) {
ALOGV("[%s] Found ready buffer", __func__);
size_t skipCount = mNewCachedSet->getLayerCount();
while (skipCount != 0) {
auto* peekThroughLayer = mNewCachedSet->getHolePunchLayer();
const size_t layerCount = currentLayerIter->getLayerCount();
for (size_t i = 0; i < layerCount; ++i) {
bool disableBlur = priorBlurLayer &&
priorBlurLayer == (*incomingLayerIter)->getOutputLayer();
OutputLayer::CompositionState& state =
(*incomingLayerIter)->getOutputLayer()->editState();
state.overrideInfo = {
.buffer = mNewCachedSet->getBuffer(),
.acquireFence = mNewCachedSet->getDrawFence(),
.displayFrame = mNewCachedSet->getTextureBounds(),
.dataspace = mNewCachedSet->getOutputDataspace(),
.displaySpace = mNewCachedSet->getOutputSpace(),
.damageRegion = Region::INVALID_REGION,
.visibleRegion = mNewCachedSet->getVisibleRegion(),
.peekThroughLayer = peekThroughLayer,
.disableBackgroundBlur = disableBlur,
};
++incomingLayerIter;
}
if (currentLayerIter->getLayerCount() > 1) {
++mInvalidatedCachedSetAges[currentLayerIter->getAge()];
}
++currentLayerIter;
skipCount -= layerCount;
}
priorBlurLayer = mNewCachedSet->getBlurLayer();
merged.emplace_back(std::move(*mNewCachedSet));
mNewCachedSet = std::nullopt;
continue;
}
}
if (!currentLayerIter->hasBufferUpdate()) {
currentLayerIter->incrementAge();
merged.emplace_back(*currentLayerIter);
// Skip the incoming layers corresponding to this valid current layer
const size_t layerCount = currentLayerIter->getLayerCount();
auto* peekThroughLayer = currentLayerIter->getHolePunchLayer();
for (size_t i = 0; i < layerCount; ++i) {
bool disableBlur =
priorBlurLayer && priorBlurLayer == (*incomingLayerIter)->getOutputLayer();
OutputLayer::CompositionState& state =
(*incomingLayerIter)->getOutputLayer()->editState();
state.overrideInfo = {
.buffer = currentLayerIter->getBuffer(),
.acquireFence = currentLayerIter->getDrawFence(),
.displayFrame = currentLayerIter->getTextureBounds(),
.dataspace = currentLayerIter->getOutputDataspace(),
.displaySpace = currentLayerIter->getOutputSpace(),
.damageRegion = Region(),
.visibleRegion = currentLayerIter->getVisibleRegion(),
.peekThroughLayer = peekThroughLayer,
.disableBackgroundBlur = disableBlur,
};
++incomingLayerIter;
}
} else if (currentLayerIter->getLayerCount() > 1) {
// Break the current layer into its constituent layers
++mInvalidatedCachedSetAges[currentLayerIter->getAge()];
for (CachedSet& layer : currentLayerIter->decompose()) {
bool disableBlur =
priorBlurLayer && priorBlurLayer == (*incomingLayerIter)->getOutputLayer();
OutputLayer::CompositionState& state =
(*incomingLayerIter)->getOutputLayer()->editState();
state.overrideInfo.disableBackgroundBlur = disableBlur;
layer.updateAge(now);
merged.emplace_back(layer);
++incomingLayerIter;
}
} else {
bool disableBlur =
priorBlurLayer && priorBlurLayer == (*incomingLayerIter)->getOutputLayer();
OutputLayer::CompositionState& state =
(*incomingLayerIter)->getOutputLayer()->editState();
state.overrideInfo.disableBackgroundBlur = disableBlur;
currentLayerIter->updateAge(now);
merged.emplace_back(*currentLayerIter);
++incomingLayerIter;
}
priorBlurLayer = currentLayerIter->getBlurLayer();
++currentLayerIter;
}
for (const CachedSet& layer : merged) {
mFlattenedDisplayCost += layer.getDisplayCost();
}
mLayers = std::move(merged);
return true;
}
std::vector<Flattener::Run> Flattener::findCandidateRuns(time_point now) const {
ATRACE_CALL();
std::vector<Run> runs;
bool isPartOfRun = false;
Run::Builder builder;
bool firstLayer = true;
bool runHasFirstLayer = false;
for (auto currentSet = mLayers.cbegin(); currentSet != mLayers.cend(); ++currentSet) {
bool layerIsInactive = now - currentSet->getLastUpdate() > mTunables.mActiveLayerTimeout;
const bool layerHasBlur = currentSet->hasBlurBehind();
const bool layerDeniedFromCaching = currentSet->cachingHintExcludesLayers();
// Layers should also be considered inactive whenever their framerate is lower than 1fps.
if (!layerIsInactive && currentSet->getLayerCount() == kNumLayersFpsConsideration) {
auto layerFps = currentSet->getFirstLayer().getState()->getFps();
if (layerFps > 0 && layerFps <= kFpsActiveThreshold) {
ATRACE_FORMAT("layer is considered inactive due to low FPS [%s] %f",
currentSet->getFirstLayer().getName().c_str(), layerFps);
layerIsInactive = true;
}
}
if (!layerDeniedFromCaching && layerIsInactive &&
(firstLayer || runHasFirstLayer || !layerHasBlur) &&
!currentSet->hasUnsupportedDataspace()) {
if (isPartOfRun) {
builder.increment();
} else {
builder.init(currentSet);
if (firstLayer) {
runHasFirstLayer = true;
}
isPartOfRun = true;
}
} else if (isPartOfRun) {
builder.setHolePunchCandidate(&(*currentSet));
// If we're here then this blur layer recently had an active buffer updating, meaning
// that there is exactly one layer. Blur radius currently is part of layer stack
// geometry, so we're also guaranteed that the background blur radius hasn't changed for
// at least as long as this new inactive cached set.
if (runHasFirstLayer && layerHasBlur &&
currentSet->getFirstLayer().getBackgroundBlurRadius() > 0) {
builder.setBlurringLayer(&(*currentSet));
}
if (auto run = builder.validateAndBuild(); run) {
runs.push_back(*run);
}
runHasFirstLayer = false;
builder.reset();
isPartOfRun = false;
}
firstLayer = false;
}
// If we're in the middle of a run at the end, we still need to validate and build it.
if (isPartOfRun) {
if (auto run = builder.validateAndBuild(); run) {
runs.push_back(*run);
}
}
ALOGV("[%s] Found %zu candidate runs", __func__, runs.size());
return runs;
}
std::optional<Flattener::Run> Flattener::findBestRun(std::vector<Flattener::Run>& runs) const {
if (runs.empty()) {
return std::nullopt;
}
// TODO (b/181192467): Choose the best run, instead of just the first.
return runs[0];
}
void Flattener::buildCachedSets(time_point now) {
ATRACE_CALL();
if (mLayers.empty()) {
ALOGV("[%s] No layers found, returning", __func__);
return;
}
// Don't try to build a new cached set if we already have a new one in progress
if (mNewCachedSet) {
return;
}
for (const CachedSet& layer : mLayers) {
// TODO (b/191997217): make it less aggressive, and sync with findCandidateRuns
if (layer.hasProtectedLayers()) {
ATRACE_NAME("layer->hasProtectedLayers()");
return;
}
}
std::vector<Run> runs = findCandidateRuns(now);
std::optional<Run> bestRun = findBestRun(runs);
if (!bestRun) {
return;
}
mNewCachedSet.emplace(*bestRun->getStart());
mNewCachedSet->setLastUpdate(now);
auto currentSet = bestRun->getStart();
while (mNewCachedSet->getLayerCount() < bestRun->getLayerLength()) {
++currentSet;
mNewCachedSet->append(*currentSet);
}
if (bestRun->getBlurringLayer()) {
mNewCachedSet->addBackgroundBlurLayer(*bestRun->getBlurringLayer());
}
if (mTunables.mEnableHolePunch && bestRun->getHolePunchCandidate() &&
bestRun->getHolePunchCandidate()->requiresHolePunch()) {
// Add the pip layer to mNewCachedSet, but in a special way - it should
// replace the buffer with a clear round rect.
mNewCachedSet->addHolePunchLayerIfFeasible(*bestRun->getHolePunchCandidate(),
bestRun->getStart() == mLayers.cbegin());
}
// TODO(b/181192467): Actually compute new LayerState vector and corresponding hash for each run
// and feedback into the predictor
++mCachedSetCreationCount;
mCachedSetCreationCost += mNewCachedSet->getCreationCost();
// note the compiler should strip the follow no-op statements when ALOGV is off
const auto dumper = [&] {
std::string setDump;
mNewCachedSet->dump(setDump);
return setDump;
};
ALOGV("[%s] Added new cached set:\n%s", __func__, dumper().c_str());
}
} // namespace android::compositionengine::impl::planner