| /* |
| * Copyright (C) 2019 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define LOG_TAG "IncrementalService" |
| |
| #include "IncrementalService.h" |
| |
| #include <android-base/logging.h> |
| #include <android-base/no_destructor.h> |
| #include <android-base/properties.h> |
| #include <android-base/stringprintf.h> |
| #include <binder/AppOpsManager.h> |
| #include <binder/Status.h> |
| #include <sys/stat.h> |
| #include <uuid/uuid.h> |
| |
| #include <charconv> |
| #include <ctime> |
| #include <iterator> |
| #include <span> |
| #include <type_traits> |
| |
| #include "IncrementalServiceValidation.h" |
| #include "Metadata.pb.h" |
| |
| using namespace std::literals; |
| |
| constexpr const char* kLoaderUsageStats = "android.permission.LOADER_USAGE_STATS"; |
| constexpr const char* kOpUsage = "android:loader_usage_stats"; |
| |
| constexpr const char* kInteractAcrossUsers = "android.permission.INTERACT_ACROSS_USERS"; |
| |
| namespace android::incremental { |
| |
| using content::pm::DataLoaderParamsParcel; |
| using content::pm::FileSystemControlParcel; |
| using content::pm::IDataLoader; |
| |
| namespace { |
| |
| using IncrementalFileSystemControlParcel = os::incremental::IncrementalFileSystemControlParcel; |
| |
| struct Constants { |
| static constexpr auto backing = "backing_store"sv; |
| static constexpr auto mount = "mount"sv; |
| static constexpr auto mountKeyPrefix = "MT_"sv; |
| static constexpr auto storagePrefix = "st"sv; |
| static constexpr auto mountpointMdPrefix = ".mountpoint."sv; |
| static constexpr auto infoMdName = ".info"sv; |
| static constexpr auto readLogsDisabledMarkerName = ".readlogs_disabled"sv; |
| static constexpr auto libDir = "lib"sv; |
| static constexpr auto libSuffix = ".so"sv; |
| static constexpr auto blockSize = 4096; |
| static constexpr auto systemPackage = "android"sv; |
| |
| static constexpr auto userStatusDelay = 100ms; |
| |
| static constexpr auto progressUpdateInterval = 1000ms; |
| static constexpr auto perUidTimeoutOffset = progressUpdateInterval * 2; |
| static constexpr auto minPerUidTimeout = progressUpdateInterval * 3; |
| |
| // If DL was up and not crashing for 10mins, we consider it healthy and reset all delays. |
| static constexpr auto healthyDataLoaderUptime = 10min; |
| |
| // For healthy DLs, we'll retry every ~5secs for ~10min |
| static constexpr auto bindRetryInterval = 5s; |
| static constexpr auto bindGracePeriod = 10min; |
| |
| static constexpr auto bindingTimeout = 1min; |
| |
| // 1s, 10s, 100s (~2min), 1000s (~15min), 10000s (~3hrs) |
| static constexpr auto minBindDelay = 1s; |
| static constexpr auto maxBindDelay = 10000s; |
| static constexpr auto bindDelayMultiplier = 10; |
| static constexpr auto bindDelayJitterDivider = 10; |
| |
| // Max interval after system invoked the DL when readlog collection can be enabled. |
| static constexpr auto readLogsMaxInterval = 2h; |
| |
| // How long should we wait till dataLoader reports destroyed. |
| static constexpr auto destroyTimeout = 10s; |
| |
| static constexpr auto anyStatus = INT_MIN; |
| }; |
| |
| static const Constants& constants() { |
| static constexpr Constants c; |
| return c; |
| } |
| |
| static bool isPageAligned(IncFsSize s) { |
| return (s & (Constants::blockSize - 1)) == 0; |
| } |
| |
| static bool getAlwaysEnableReadTimeoutsForSystemDataLoaders() { |
| return android::base:: |
| GetBoolProperty("debug.incremental.always_enable_read_timeouts_for_system_dataloaders", |
| true); |
| } |
| |
| static bool getEnableReadTimeoutsAfterInstall() { |
| return android::base::GetBoolProperty("debug.incremental.enable_read_timeouts_after_install", |
| true); |
| } |
| |
| static bool getEnforceReadLogsMaxIntervalForSystemDataLoaders() { |
| return android::base::GetBoolProperty("debug.incremental.enforce_readlogs_max_interval_for_" |
| "system_dataloaders", |
| false); |
| } |
| |
| static Seconds getReadLogsMaxInterval() { |
| constexpr int limit = duration_cast<Seconds>(Constants::readLogsMaxInterval).count(); |
| int readlogs_max_interval_secs = |
| std::min(limit, |
| android::base::GetIntProperty< |
| int>("debug.incremental.readlogs_max_interval_sec", limit)); |
| return Seconds{readlogs_max_interval_secs}; |
| } |
| |
| template <base::LogSeverity level = base::ERROR> |
| bool mkdirOrLog(std::string_view name, int mode = 0770, bool allowExisting = true) { |
| auto cstr = path::c_str(name); |
| if (::mkdir(cstr, mode)) { |
| if (!allowExisting || errno != EEXIST) { |
| PLOG(level) << "Can't create directory '" << name << '\''; |
| return false; |
| } |
| struct stat st; |
| if (::stat(cstr, &st) || !S_ISDIR(st.st_mode)) { |
| PLOG(level) << "Path exists but is not a directory: '" << name << '\''; |
| return false; |
| } |
| } |
| if (::chmod(cstr, mode)) { |
| PLOG(level) << "Changing permission failed for '" << name << '\''; |
| return false; |
| } |
| |
| return true; |
| } |
| |
| static std::string toMountKey(std::string_view path) { |
| if (path.empty()) { |
| return "@none"; |
| } |
| if (path == "/"sv) { |
| return "@root"; |
| } |
| if (path::isAbsolute(path)) { |
| path.remove_prefix(1); |
| } |
| if (path.size() > 16) { |
| path = path.substr(0, 16); |
| } |
| std::string res(path); |
| std::replace_if( |
| res.begin(), res.end(), [](char c) { return c == '/' || c == '@'; }, '_'); |
| return std::string(constants().mountKeyPrefix) += res; |
| } |
| |
| static std::pair<std::string, std::string> makeMountDir(std::string_view incrementalDir, |
| std::string_view path) { |
| auto mountKey = toMountKey(path); |
| const auto prefixSize = mountKey.size(); |
| for (int counter = 0; counter < 1000; |
| mountKey.resize(prefixSize), base::StringAppendF(&mountKey, "%d", counter++)) { |
| auto mountRoot = path::join(incrementalDir, mountKey); |
| if (mkdirOrLog(mountRoot, 0777, false)) { |
| return {mountKey, mountRoot}; |
| } |
| } |
| return {}; |
| } |
| |
| template <class Map> |
| typename Map::const_iterator findParentPath(const Map& map, std::string_view path) { |
| const auto nextIt = map.upper_bound(path); |
| if (nextIt == map.begin()) { |
| return map.end(); |
| } |
| const auto suspectIt = std::prev(nextIt); |
| if (!path::startsWith(path, suspectIt->first)) { |
| return map.end(); |
| } |
| return suspectIt; |
| } |
| |
| static base::unique_fd dup(base::borrowed_fd fd) { |
| const auto res = fcntl(fd.get(), F_DUPFD_CLOEXEC, 0); |
| return base::unique_fd(res); |
| } |
| |
| template <class ProtoMessage, class Control> |
| static ProtoMessage parseFromIncfs(const IncFsWrapper* incfs, const Control& control, |
| std::string_view path) { |
| auto md = incfs->getMetadata(control, path); |
| ProtoMessage message; |
| return message.ParseFromArray(md.data(), md.size()) ? message : ProtoMessage{}; |
| } |
| |
| static bool isValidMountTarget(std::string_view path) { |
| return path::isAbsolute(path) && path::isEmptyDir(path).value_or(true); |
| } |
| |
| std::string makeUniqueName(std::string_view prefix) { |
| static constexpr auto uuidStringSize = 36; |
| |
| uuid_t guid; |
| uuid_generate(guid); |
| |
| std::string name; |
| const auto prefixSize = prefix.size(); |
| name.reserve(prefixSize + uuidStringSize); |
| |
| name = prefix; |
| name.resize(prefixSize + uuidStringSize); |
| uuid_unparse(guid, name.data() + prefixSize); |
| |
| return name; |
| } |
| |
| std::string makeBindMdName() { |
| return makeUniqueName(constants().mountpointMdPrefix); |
| } |
| |
| static bool checkReadLogsDisabledMarker(std::string_view root) { |
| const auto markerPath = path::c_str(path::join(root, constants().readLogsDisabledMarkerName)); |
| struct stat st; |
| return (::stat(markerPath, &st) == 0); |
| } |
| |
| } // namespace |
| |
| IncrementalService::IncFsMount::~IncFsMount() { |
| if (dataLoaderStub) { |
| dataLoaderStub->cleanupResources(); |
| dataLoaderStub = {}; |
| } |
| control.close(); |
| LOG(INFO) << "Unmounting and cleaning up mount " << mountId << " with root '" << root << '\''; |
| for (auto&& [target, _] : bindPoints) { |
| LOG(INFO) << " bind: " << target; |
| incrementalService.mVold->unmountIncFs(target); |
| } |
| LOG(INFO) << " root: " << root; |
| incrementalService.mVold->unmountIncFs(path::join(root, constants().mount)); |
| cleanupFilesystem(root); |
| } |
| |
| auto IncrementalService::IncFsMount::makeStorage(StorageId id) -> StorageMap::iterator { |
| std::string name; |
| for (int no = nextStorageDirNo.fetch_add(1, std::memory_order_relaxed), i = 0; |
| i < 1024 && no >= 0; no = nextStorageDirNo.fetch_add(1, std::memory_order_relaxed), ++i) { |
| name.clear(); |
| base::StringAppendF(&name, "%.*s_%d_%d", int(constants().storagePrefix.size()), |
| constants().storagePrefix.data(), id, no); |
| auto fullName = path::join(root, constants().mount, name); |
| if (auto err = incrementalService.mIncFs->makeDir(control, fullName, 0755); !err) { |
| std::lock_guard l(lock); |
| return storages.insert_or_assign(id, Storage{std::move(fullName)}).first; |
| } else if (err != EEXIST) { |
| LOG(ERROR) << __func__ << "(): failed to create dir |" << fullName << "| " << err; |
| break; |
| } |
| } |
| nextStorageDirNo = 0; |
| return storages.end(); |
| } |
| |
| template <class Func> |
| static auto makeCleanup(Func&& f) requires(!std::is_lvalue_reference_v<Func>) { |
| // ok to move a 'forwarding' reference here as lvalues are disabled anyway |
| auto deleter = [f = std::move(f)](auto) { // NOLINT |
| f(); |
| }; |
| // &f is a dangling pointer here, but we actually never use it as deleter moves it in. |
| return std::unique_ptr<Func, decltype(deleter)>(&f, std::move(deleter)); |
| } |
| |
| static auto openDir(const char* dir) { |
| struct DirCloser { |
| void operator()(DIR* d) const noexcept { ::closedir(d); } |
| }; |
| return std::unique_ptr<DIR, DirCloser>(::opendir(dir)); |
| } |
| |
| static auto openDir(std::string_view dir) { |
| return openDir(path::c_str(dir)); |
| } |
| |
| static int rmDirContent(const char* path) { |
| auto dir = openDir(path); |
| if (!dir) { |
| return -EINVAL; |
| } |
| while (auto entry = ::readdir(dir.get())) { |
| if (entry->d_name == "."sv || entry->d_name == ".."sv) { |
| continue; |
| } |
| auto fullPath = base::StringPrintf("%s/%s", path, entry->d_name); |
| if (entry->d_type == DT_DIR) { |
| if (const auto err = rmDirContent(fullPath.c_str()); err != 0) { |
| PLOG(WARNING) << "Failed to delete " << fullPath << " content"; |
| return err; |
| } |
| if (const auto err = ::rmdir(fullPath.c_str()); err != 0) { |
| PLOG(WARNING) << "Failed to rmdir " << fullPath; |
| return err; |
| } |
| } else { |
| if (const auto err = ::unlink(fullPath.c_str()); err != 0) { |
| PLOG(WARNING) << "Failed to delete " << fullPath; |
| return err; |
| } |
| } |
| } |
| return 0; |
| } |
| |
| void IncrementalService::IncFsMount::cleanupFilesystem(std::string_view root) { |
| rmDirContent(path::join(root, constants().backing).c_str()); |
| ::rmdir(path::join(root, constants().backing).c_str()); |
| ::rmdir(path::join(root, constants().mount).c_str()); |
| ::rmdir(path::c_str(root)); |
| } |
| |
| void IncrementalService::IncFsMount::setFlag(StorageFlags flag, bool value) { |
| if (value) { |
| flags |= flag; |
| } else { |
| flags &= ~flag; |
| } |
| } |
| |
| IncrementalService::IncrementalService(ServiceManagerWrapper&& sm, std::string_view rootDir) |
| : mVold(sm.getVoldService()), |
| mDataLoaderManager(sm.getDataLoaderManager()), |
| mIncFs(sm.getIncFs()), |
| mAppOpsManager(sm.getAppOpsManager()), |
| mJni(sm.getJni()), |
| mLooper(sm.getLooper()), |
| mTimedQueue(sm.getTimedQueue()), |
| mProgressUpdateJobQueue(sm.getProgressUpdateJobQueue()), |
| mFs(sm.getFs()), |
| mClock(sm.getClock()), |
| mIncrementalDir(rootDir) { |
| CHECK(mVold) << "Vold service is unavailable"; |
| CHECK(mDataLoaderManager) << "DataLoaderManagerService is unavailable"; |
| CHECK(mAppOpsManager) << "AppOpsManager is unavailable"; |
| CHECK(mJni) << "JNI is unavailable"; |
| CHECK(mLooper) << "Looper is unavailable"; |
| CHECK(mTimedQueue) << "TimedQueue is unavailable"; |
| CHECK(mProgressUpdateJobQueue) << "mProgressUpdateJobQueue is unavailable"; |
| CHECK(mFs) << "Fs is unavailable"; |
| CHECK(mClock) << "Clock is unavailable"; |
| |
| mJobQueue.reserve(16); |
| mJobProcessor = std::thread([this]() { |
| mJni->initializeForCurrentThread(); |
| runJobProcessing(); |
| }); |
| mCmdLooperThread = std::thread([this]() { |
| mJni->initializeForCurrentThread(); |
| runCmdLooper(); |
| }); |
| |
| const auto mountedRootNames = adoptMountedInstances(); |
| mountExistingImages(mountedRootNames); |
| } |
| |
| IncrementalService::~IncrementalService() { |
| { |
| std::lock_guard lock(mJobMutex); |
| mRunning = false; |
| } |
| mJobCondition.notify_all(); |
| mJobProcessor.join(); |
| mLooper->wake(); |
| mCmdLooperThread.join(); |
| mTimedQueue->stop(); |
| mProgressUpdateJobQueue->stop(); |
| // Ensure that mounts are destroyed while the service is still valid. |
| mBindsByPath.clear(); |
| mMounts.clear(); |
| } |
| |
| static const char* toString(IncrementalService::BindKind kind) { |
| switch (kind) { |
| case IncrementalService::BindKind::Temporary: |
| return "Temporary"; |
| case IncrementalService::BindKind::Permanent: |
| return "Permanent"; |
| } |
| } |
| |
| template <class Duration> |
| static int64_t elapsedMcs(Duration start, Duration end) { |
| return std::chrono::duration_cast<std::chrono::microseconds>(end - start).count(); |
| } |
| |
| int64_t IncrementalService::elapsedUsSinceMonoTs(uint64_t monoTsUs) { |
| const auto now = mClock->now(); |
| const auto nowUs = static_cast<uint64_t>( |
| duration_cast<std::chrono::microseconds>(now.time_since_epoch()).count()); |
| return nowUs - monoTsUs; |
| } |
| |
| static const char* loadingStateToString(incfs::LoadingState state) { |
| switch (state) { |
| case (incfs::LoadingState::Full): |
| return "Full"; |
| case (incfs::LoadingState::MissingBlocks): |
| return "MissingBlocks"; |
| default: |
| return "error obtaining loading state"; |
| } |
| } |
| |
| void IncrementalService::onDump(int fd) { |
| dprintf(fd, "Incremental is %s\n", incfs::enabled() ? "ENABLED" : "DISABLED"); |
| dprintf(fd, "IncFs features: 0x%x\n", int(mIncFs->features())); |
| dprintf(fd, "Incremental dir: %s\n", mIncrementalDir.c_str()); |
| |
| std::unique_lock l(mLock); |
| |
| dprintf(fd, "Mounts (%d): {\n", int(mMounts.size())); |
| for (auto&& [id, ifs] : mMounts) { |
| std::unique_lock ll(ifs->lock); |
| const IncFsMount& mnt = *ifs; |
| dprintf(fd, " [%d]: {\n", id); |
| if (id != mnt.mountId) { |
| dprintf(fd, " reference to mountId: %d\n", mnt.mountId); |
| } else { |
| dprintf(fd, " mountId: %d\n", mnt.mountId); |
| dprintf(fd, " root: %s\n", mnt.root.c_str()); |
| const auto& metricsInstanceName = ifs->metricsKey; |
| dprintf(fd, " metrics instance name: %s\n", path::c_str(metricsInstanceName).get()); |
| dprintf(fd, " nextStorageDirNo: %d\n", mnt.nextStorageDirNo.load()); |
| dprintf(fd, " flags: %d\n", int(mnt.flags)); |
| if (mnt.startLoadingTs.time_since_epoch() == Clock::duration::zero()) { |
| dprintf(fd, " not loading\n"); |
| } else { |
| dprintf(fd, " startLoading: %llds\n", |
| (long long)(elapsedMcs(mnt.startLoadingTs, Clock::now()) / 1000000)); |
| } |
| if (mnt.dataLoaderStub) { |
| mnt.dataLoaderStub->onDump(fd); |
| } else { |
| dprintf(fd, " dataLoader: null\n"); |
| } |
| dprintf(fd, " storages (%d): {\n", int(mnt.storages.size())); |
| for (auto&& [storageId, storage] : mnt.storages) { |
| auto&& ifs = getIfsLocked(storageId); |
| dprintf(fd, " [%d] -> [%s] (%d %% loaded)(%s) \n", storageId, |
| storage.name.c_str(), |
| (int)(getLoadingProgressFromPath(mnt, storage.name.c_str()).getProgress() * |
| 100), |
| ifs ? loadingStateToString(mIncFs->isEverythingFullyLoaded(ifs->control)) |
| : "error obtaining ifs"); |
| } |
| dprintf(fd, " }\n"); |
| |
| dprintf(fd, " bindPoints (%d): {\n", int(mnt.bindPoints.size())); |
| for (auto&& [target, bind] : mnt.bindPoints) { |
| dprintf(fd, " [%s]->[%d]:\n", target.c_str(), bind.storage); |
| dprintf(fd, " savedFilename: %s\n", bind.savedFilename.c_str()); |
| dprintf(fd, " sourceDir: %s\n", bind.sourceDir.c_str()); |
| dprintf(fd, " kind: %s\n", toString(bind.kind)); |
| } |
| dprintf(fd, " }\n"); |
| |
| dprintf(fd, " incfsMetrics: {\n"); |
| const auto incfsMetrics = mIncFs->getMetrics(metricsInstanceName); |
| if (incfsMetrics) { |
| dprintf(fd, " readsDelayedMin: %d\n", incfsMetrics.value().readsDelayedMin); |
| dprintf(fd, " readsDelayedMinUs: %lld\n", |
| (long long)incfsMetrics.value().readsDelayedMinUs); |
| dprintf(fd, " readsDelayedPending: %d\n", |
| incfsMetrics.value().readsDelayedPending); |
| dprintf(fd, " readsDelayedPendingUs: %lld\n", |
| (long long)incfsMetrics.value().readsDelayedPendingUs); |
| dprintf(fd, " readsFailedHashVerification: %d\n", |
| incfsMetrics.value().readsFailedHashVerification); |
| dprintf(fd, " readsFailedOther: %d\n", incfsMetrics.value().readsFailedOther); |
| dprintf(fd, " readsFailedTimedOut: %d\n", |
| incfsMetrics.value().readsFailedTimedOut); |
| } else { |
| dprintf(fd, " Metrics not available. Errno: %d\n", errno); |
| } |
| dprintf(fd, " }\n"); |
| |
| const auto lastReadError = mIncFs->getLastReadError(ifs->control); |
| const auto errorNo = errno; |
| dprintf(fd, " lastReadError: {\n"); |
| if (lastReadError) { |
| if (lastReadError->timestampUs == 0) { |
| dprintf(fd, " No read errors.\n"); |
| } else { |
| dprintf(fd, " fileId: %s\n", |
| IncFsWrapper::toString(lastReadError->id).c_str()); |
| dprintf(fd, " time: %llu microseconds ago\n", |
| (unsigned long long)elapsedUsSinceMonoTs(lastReadError->timestampUs)); |
| dprintf(fd, " blockIndex: %d\n", lastReadError->block); |
| dprintf(fd, " errno: %d\n", lastReadError->errorNo); |
| } |
| } else { |
| dprintf(fd, " Info not available. Errno: %d\n", errorNo); |
| } |
| dprintf(fd, " }\n"); |
| } |
| dprintf(fd, " }\n"); |
| } |
| dprintf(fd, "}\n"); |
| dprintf(fd, "Sorted binds (%d): {\n", int(mBindsByPath.size())); |
| for (auto&& [target, mountPairIt] : mBindsByPath) { |
| const auto& bind = mountPairIt->second; |
| dprintf(fd, " [%s]->[%d]:\n", target.c_str(), bind.storage); |
| dprintf(fd, " savedFilename: %s\n", bind.savedFilename.c_str()); |
| dprintf(fd, " sourceDir: %s\n", bind.sourceDir.c_str()); |
| dprintf(fd, " kind: %s\n", toString(bind.kind)); |
| } |
| dprintf(fd, "}\n"); |
| } |
| |
| bool IncrementalService::needStartDataLoaderLocked(IncFsMount& ifs) { |
| if (!ifs.dataLoaderStub) { |
| return false; |
| } |
| if (ifs.dataLoaderStub->isSystemDataLoader()) { |
| return true; |
| } |
| |
| return mIncFs->isEverythingFullyLoaded(ifs.control) == incfs::LoadingState::MissingBlocks; |
| } |
| |
| void IncrementalService::onSystemReady() { |
| if (mSystemReady.exchange(true)) { |
| return; |
| } |
| |
| std::vector<IfsMountPtr> mounts; |
| { |
| std::lock_guard l(mLock); |
| mounts.reserve(mMounts.size()); |
| for (auto&& [id, ifs] : mMounts) { |
| std::unique_lock ll(ifs->lock); |
| |
| if (ifs->mountId != id) { |
| continue; |
| } |
| |
| if (needStartDataLoaderLocked(*ifs)) { |
| mounts.push_back(ifs); |
| } |
| } |
| } |
| |
| if (mounts.empty()) { |
| return; |
| } |
| |
| std::thread([this, mounts = std::move(mounts)]() { |
| mJni->initializeForCurrentThread(); |
| for (auto&& ifs : mounts) { |
| std::unique_lock l(ifs->lock); |
| if (ifs->dataLoaderStub) { |
| ifs->dataLoaderStub->requestStart(); |
| } |
| } |
| }).detach(); |
| } |
| |
| auto IncrementalService::getStorageSlotLocked() -> MountMap::iterator { |
| for (;;) { |
| if (mNextId == kMaxStorageId) { |
| mNextId = 0; |
| } |
| auto id = ++mNextId; |
| auto [it, inserted] = mMounts.try_emplace(id, nullptr); |
| if (inserted) { |
| return it; |
| } |
| } |
| } |
| |
| StorageId IncrementalService::createStorage(std::string_view mountPoint, |
| content::pm::DataLoaderParamsParcel dataLoaderParams, |
| CreateOptions options) { |
| LOG(INFO) << "createStorage: " << mountPoint << " | " << int(options); |
| if (!path::isAbsolute(mountPoint)) { |
| LOG(ERROR) << "path is not absolute: " << mountPoint; |
| return kInvalidStorageId; |
| } |
| |
| auto mountNorm = path::normalize(mountPoint); |
| { |
| const auto id = findStorageId(mountNorm); |
| if (id != kInvalidStorageId) { |
| if (options & CreateOptions::OpenExisting) { |
| LOG(INFO) << "Opened existing storage " << id; |
| return id; |
| } |
| LOG(ERROR) << "Directory " << mountPoint << " is already mounted at storage " << id; |
| return kInvalidStorageId; |
| } |
| } |
| |
| if (!(options & CreateOptions::CreateNew)) { |
| LOG(ERROR) << "not requirested create new storage, and it doesn't exist: " << mountPoint; |
| return kInvalidStorageId; |
| } |
| |
| if (!path::isEmptyDir(mountNorm)) { |
| LOG(ERROR) << "Mounting over existing non-empty directory is not supported: " << mountNorm; |
| return kInvalidStorageId; |
| } |
| auto [mountKey, mountRoot] = makeMountDir(mIncrementalDir, mountNorm); |
| if (mountRoot.empty()) { |
| LOG(ERROR) << "Bad mount point"; |
| return kInvalidStorageId; |
| } |
| // Make sure the code removes all crap it may create while still failing. |
| auto firstCleanup = [](const std::string* ptr) { IncFsMount::cleanupFilesystem(*ptr); }; |
| auto firstCleanupOnFailure = |
| std::unique_ptr<std::string, decltype(firstCleanup)>(&mountRoot, firstCleanup); |
| |
| auto mountTarget = path::join(mountRoot, constants().mount); |
| const auto backing = path::join(mountRoot, constants().backing); |
| if (!mkdirOrLog(backing, 0777) || !mkdirOrLog(mountTarget)) { |
| return kInvalidStorageId; |
| } |
| |
| std::string metricsKey; |
| IncFsMount::Control control; |
| { |
| std::lock_guard l(mMountOperationLock); |
| IncrementalFileSystemControlParcel controlParcel; |
| |
| if (auto err = rmDirContent(backing.c_str())) { |
| LOG(ERROR) << "Coudn't clean the backing directory " << backing << ": " << err; |
| return kInvalidStorageId; |
| } |
| if (!mkdirOrLog(path::join(backing, ".index"), 0777)) { |
| return kInvalidStorageId; |
| } |
| if (!mkdirOrLog(path::join(backing, ".incomplete"), 0777)) { |
| return kInvalidStorageId; |
| } |
| metricsKey = makeUniqueName(mountKey); |
| auto status = mVold->mountIncFs(backing, mountTarget, 0, metricsKey, &controlParcel); |
| if (!status.isOk()) { |
| LOG(ERROR) << "Vold::mountIncFs() failed: " << status.toString8(); |
| return kInvalidStorageId; |
| } |
| if (controlParcel.cmd.get() < 0 || controlParcel.pendingReads.get() < 0 || |
| controlParcel.log.get() < 0) { |
| LOG(ERROR) << "Vold::mountIncFs() returned invalid control parcel."; |
| return kInvalidStorageId; |
| } |
| int cmd = controlParcel.cmd.release().release(); |
| int pendingReads = controlParcel.pendingReads.release().release(); |
| int logs = controlParcel.log.release().release(); |
| int blocksWritten = |
| controlParcel.blocksWritten ? controlParcel.blocksWritten->release().release() : -1; |
| control = mIncFs->createControl(cmd, pendingReads, logs, blocksWritten); |
| } |
| |
| std::unique_lock l(mLock); |
| const auto mountIt = getStorageSlotLocked(); |
| const auto mountId = mountIt->first; |
| l.unlock(); |
| |
| auto ifs = std::make_shared<IncFsMount>(std::move(mountRoot), std::move(metricsKey), mountId, |
| std::move(control), *this); |
| // Now it's the |ifs|'s responsibility to clean up after itself, and the only cleanup we need |
| // is the removal of the |ifs|. |
| (void)firstCleanupOnFailure.release(); |
| |
| auto secondCleanup = [this, &l](auto itPtr) { |
| if (!l.owns_lock()) { |
| l.lock(); |
| } |
| mMounts.erase(*itPtr); |
| }; |
| auto secondCleanupOnFailure = |
| std::unique_ptr<decltype(mountIt), decltype(secondCleanup)>(&mountIt, secondCleanup); |
| |
| const auto storageIt = ifs->makeStorage(ifs->mountId); |
| if (storageIt == ifs->storages.end()) { |
| LOG(ERROR) << "Can't create a default storage directory"; |
| return kInvalidStorageId; |
| } |
| |
| { |
| metadata::Mount m; |
| m.mutable_storage()->set_id(ifs->mountId); |
| m.mutable_loader()->set_type((int)dataLoaderParams.type); |
| m.mutable_loader()->set_package_name(std::move(dataLoaderParams.packageName)); |
| m.mutable_loader()->set_class_name(std::move(dataLoaderParams.className)); |
| m.mutable_loader()->set_arguments(std::move(dataLoaderParams.arguments)); |
| const auto metadata = m.SerializeAsString(); |
| if (auto err = |
| mIncFs->makeFile(ifs->control, |
| path::join(ifs->root, constants().mount, |
| constants().infoMdName), |
| 0777, idFromMetadata(metadata), |
| {.metadata = {metadata.data(), (IncFsSize)metadata.size()}})) { |
| LOG(ERROR) << "Saving mount metadata failed: " << -err; |
| return kInvalidStorageId; |
| } |
| } |
| |
| const auto bk = |
| (options & CreateOptions::PermanentBind) ? BindKind::Permanent : BindKind::Temporary; |
| if (auto err = addBindMount(*ifs, storageIt->first, storageIt->second.name, |
| std::string(storageIt->second.name), std::move(mountNorm), bk, l); |
| err < 0) { |
| LOG(ERROR) << "Adding bind mount failed: " << -err; |
| return kInvalidStorageId; |
| } |
| |
| // Done here as well, all data structures are in good state. |
| (void)secondCleanupOnFailure.release(); |
| |
| mountIt->second = std::move(ifs); |
| l.unlock(); |
| |
| LOG(INFO) << "created storage " << mountId; |
| return mountId; |
| } |
| |
| StorageId IncrementalService::createLinkedStorage(std::string_view mountPoint, |
| StorageId linkedStorage, |
| IncrementalService::CreateOptions options) { |
| if (!isValidMountTarget(mountPoint)) { |
| LOG(ERROR) << "Mount point is invalid or missing"; |
| return kInvalidStorageId; |
| } |
| |
| std::unique_lock l(mLock); |
| auto ifs = getIfsLocked(linkedStorage); |
| if (!ifs) { |
| LOG(ERROR) << "Ifs unavailable"; |
| return kInvalidStorageId; |
| } |
| |
| const auto mountIt = getStorageSlotLocked(); |
| const auto storageId = mountIt->first; |
| const auto storageIt = ifs->makeStorage(storageId); |
| if (storageIt == ifs->storages.end()) { |
| LOG(ERROR) << "Can't create a new storage"; |
| mMounts.erase(mountIt); |
| return kInvalidStorageId; |
| } |
| |
| l.unlock(); |
| |
| const auto bk = |
| (options & CreateOptions::PermanentBind) ? BindKind::Permanent : BindKind::Temporary; |
| if (auto err = addBindMount(*ifs, storageIt->first, storageIt->second.name, |
| std::string(storageIt->second.name), path::normalize(mountPoint), |
| bk, l); |
| err < 0) { |
| LOG(ERROR) << "bindMount failed with error: " << err; |
| (void)mIncFs->unlink(ifs->control, storageIt->second.name); |
| ifs->storages.erase(storageIt); |
| return kInvalidStorageId; |
| } |
| |
| mountIt->second = ifs; |
| return storageId; |
| } |
| |
| bool IncrementalService::startLoading(StorageId storageId, |
| content::pm::DataLoaderParamsParcel dataLoaderParams, |
| DataLoaderStatusListener statusListener, |
| const StorageHealthCheckParams& healthCheckParams, |
| StorageHealthListener healthListener, |
| std::vector<PerUidReadTimeouts> perUidReadTimeouts) { |
| // Per Uid timeouts. |
| if (!perUidReadTimeouts.empty()) { |
| setUidReadTimeouts(storageId, std::move(perUidReadTimeouts)); |
| } |
| |
| IfsMountPtr ifs; |
| DataLoaderStubPtr dataLoaderStub; |
| |
| // Re-initialize DataLoader. |
| { |
| ifs = getIfs(storageId); |
| if (!ifs) { |
| return false; |
| } |
| |
| std::unique_lock l(ifs->lock); |
| dataLoaderStub = std::exchange(ifs->dataLoaderStub, nullptr); |
| } |
| |
| if (dataLoaderStub) { |
| dataLoaderStub->cleanupResources(); |
| dataLoaderStub = {}; |
| } |
| |
| { |
| std::unique_lock l(ifs->lock); |
| if (ifs->dataLoaderStub) { |
| LOG(INFO) << "Skipped data loader stub creation because it already exists"; |
| return false; |
| } |
| |
| prepareDataLoaderLocked(*ifs, std::move(dataLoaderParams), std::move(statusListener), |
| healthCheckParams, std::move(healthListener)); |
| CHECK(ifs->dataLoaderStub); |
| dataLoaderStub = ifs->dataLoaderStub; |
| |
| // Disable long read timeouts for non-system dataloaders. |
| // To be re-enabled after installation is complete. |
| ifs->setReadTimeoutsRequested(dataLoaderStub->isSystemDataLoader() && |
| getAlwaysEnableReadTimeoutsForSystemDataLoaders()); |
| applyStorageParamsLocked(*ifs); |
| } |
| |
| if (dataLoaderStub->isSystemDataLoader() && |
| !getEnforceReadLogsMaxIntervalForSystemDataLoaders()) { |
| // Readlogs from system dataloader (adb) can always be collected. |
| ifs->startLoadingTs = TimePoint::max(); |
| } else { |
| // Assign time when installation wants the DL to start streaming. |
| const auto startLoadingTs = mClock->now(); |
| ifs->startLoadingTs = startLoadingTs; |
| // Setup a callback to disable the readlogs after max interval. |
| addTimedJob(*mTimedQueue, storageId, getReadLogsMaxInterval(), |
| [this, storageId, startLoadingTs]() { |
| const auto ifs = getIfs(storageId); |
| if (!ifs) { |
| LOG(WARNING) << "Can't disable the readlogs, invalid storageId: " |
| << storageId; |
| return; |
| } |
| std::unique_lock l(ifs->lock); |
| if (ifs->startLoadingTs != startLoadingTs) { |
| LOG(INFO) << "Can't disable the readlogs, timestamp mismatch (new " |
| "installation?): " |
| << storageId; |
| return; |
| } |
| disableReadLogsLocked(*ifs); |
| }); |
| } |
| |
| return dataLoaderStub->requestStart(); |
| } |
| |
| void IncrementalService::onInstallationComplete(StorageId storage) { |
| IfsMountPtr ifs = getIfs(storage); |
| if (!ifs) { |
| return; |
| } |
| |
| // Always enable long read timeouts after installation is complete. |
| std::unique_lock l(ifs->lock); |
| ifs->setReadTimeoutsRequested(getEnableReadTimeoutsAfterInstall()); |
| applyStorageParamsLocked(*ifs); |
| } |
| |
| IncrementalService::BindPathMap::const_iterator IncrementalService::findStorageLocked( |
| std::string_view path) const { |
| return findParentPath(mBindsByPath, path); |
| } |
| |
| StorageId IncrementalService::findStorageId(std::string_view path) const { |
| std::lock_guard l(mLock); |
| auto it = findStorageLocked(path); |
| if (it == mBindsByPath.end()) { |
| return kInvalidStorageId; |
| } |
| return it->second->second.storage; |
| } |
| |
| void IncrementalService::disallowReadLogs(StorageId storageId) { |
| const auto ifs = getIfs(storageId); |
| if (!ifs) { |
| LOG(ERROR) << "disallowReadLogs failed, invalid storageId: " << storageId; |
| return; |
| } |
| |
| std::unique_lock l(ifs->lock); |
| if (!ifs->readLogsAllowed()) { |
| return; |
| } |
| ifs->disallowReadLogs(); |
| |
| const auto metadata = constants().readLogsDisabledMarkerName; |
| if (auto err = mIncFs->makeFile(ifs->control, |
| path::join(ifs->root, constants().mount, |
| constants().readLogsDisabledMarkerName), |
| 0777, idFromMetadata(metadata), {})) { |
| //{.metadata = {metadata.data(), (IncFsSize)metadata.size()}})) { |
| LOG(ERROR) << "Failed to make marker file for storageId: " << storageId << " err: " << -err; |
| return; |
| } |
| |
| disableReadLogsLocked(*ifs); |
| } |
| |
| int IncrementalService::setStorageParams(StorageId storageId, bool enableReadLogs) { |
| const auto ifs = getIfs(storageId); |
| if (!ifs) { |
| LOG(ERROR) << "setStorageParams failed, invalid storageId: " << storageId; |
| return -EINVAL; |
| } |
| |
| std::string packageName; |
| |
| { |
| std::unique_lock l(ifs->lock); |
| if (!enableReadLogs) { |
| return disableReadLogsLocked(*ifs); |
| } |
| |
| if (!ifs->readLogsAllowed()) { |
| LOG(ERROR) << "enableReadLogs failed, readlogs disallowed for storageId: " << storageId; |
| return -EPERM; |
| } |
| |
| if (!ifs->dataLoaderStub) { |
| // This should never happen - only DL can call enableReadLogs. |
| LOG(ERROR) << "enableReadLogs failed: invalid state"; |
| return -EPERM; |
| } |
| |
| // Check installation time. |
| const auto now = mClock->now(); |
| const auto startLoadingTs = ifs->startLoadingTs; |
| if (startLoadingTs <= now && now - startLoadingTs > getReadLogsMaxInterval()) { |
| LOG(ERROR) |
| << "enableReadLogs failed, readlogs can't be enabled at this time, storageId: " |
| << storageId; |
| return -EPERM; |
| } |
| |
| packageName = ifs->dataLoaderStub->params().packageName; |
| ifs->setReadLogsRequested(true); |
| } |
| |
| // Check loader usage stats permission and apop. |
| if (auto status = |
| mAppOpsManager->checkPermission(kLoaderUsageStats, kOpUsage, packageName.c_str()); |
| !status.isOk()) { |
| LOG(ERROR) << " Permission: " << kLoaderUsageStats |
| << " check failed: " << status.toString8(); |
| return fromBinderStatus(status); |
| } |
| |
| // Check multiuser permission. |
| if (auto status = |
| mAppOpsManager->checkPermission(kInteractAcrossUsers, nullptr, packageName.c_str()); |
| !status.isOk()) { |
| LOG(ERROR) << " Permission: " << kInteractAcrossUsers |
| << " check failed: " << status.toString8(); |
| return fromBinderStatus(status); |
| } |
| |
| { |
| std::unique_lock l(ifs->lock); |
| if (!ifs->readLogsRequested()) { |
| return 0; |
| } |
| if (auto status = applyStorageParamsLocked(*ifs); status != 0) { |
| return status; |
| } |
| } |
| |
| registerAppOpsCallback(packageName); |
| |
| return 0; |
| } |
| |
| int IncrementalService::disableReadLogsLocked(IncFsMount& ifs) { |
| ifs.setReadLogsRequested(false); |
| return applyStorageParamsLocked(ifs); |
| } |
| |
| int IncrementalService::applyStorageParamsLocked(IncFsMount& ifs) { |
| os::incremental::IncrementalFileSystemControlParcel control; |
| control.cmd.reset(dup(ifs.control.cmd())); |
| control.pendingReads.reset(dup(ifs.control.pendingReads())); |
| auto logsFd = ifs.control.logs(); |
| if (logsFd >= 0) { |
| control.log.reset(dup(logsFd)); |
| } |
| |
| bool enableReadLogs = ifs.readLogsRequested(); |
| bool enableReadTimeouts = ifs.readTimeoutsRequested(); |
| |
| std::lock_guard l(mMountOperationLock); |
| auto status = mVold->setIncFsMountOptions(control, enableReadLogs, enableReadTimeouts, |
| ifs.metricsKey); |
| if (status.isOk()) { |
| // Store states. |
| ifs.setReadLogsEnabled(enableReadLogs); |
| ifs.setReadTimeoutsEnabled(enableReadTimeouts); |
| } else { |
| LOG(ERROR) << "applyStorageParams failed: " << status.toString8(); |
| } |
| return status.isOk() ? 0 : fromBinderStatus(status); |
| } |
| |
| void IncrementalService::deleteStorage(StorageId storageId) { |
| const auto ifs = getIfs(storageId); |
| if (!ifs) { |
| return; |
| } |
| deleteStorage(*ifs); |
| } |
| |
| void IncrementalService::deleteStorage(IncrementalService::IncFsMount& ifs) { |
| std::unique_lock l(ifs.lock); |
| deleteStorageLocked(ifs, std::move(l)); |
| } |
| |
| void IncrementalService::deleteStorageLocked(IncrementalService::IncFsMount& ifs, |
| std::unique_lock<std::mutex>&& ifsLock) { |
| const auto storages = std::move(ifs.storages); |
| // Don't move the bind points out: Ifs's dtor will use them to unmount everything. |
| const auto bindPoints = ifs.bindPoints; |
| ifsLock.unlock(); |
| |
| std::lock_guard l(mLock); |
| for (auto&& [id, _] : storages) { |
| if (id != ifs.mountId) { |
| mMounts.erase(id); |
| } |
| } |
| for (auto&& [path, _] : bindPoints) { |
| mBindsByPath.erase(path); |
| } |
| mMounts.erase(ifs.mountId); |
| } |
| |
| StorageId IncrementalService::openStorage(std::string_view pathInMount) { |
| if (!path::isAbsolute(pathInMount)) { |
| return kInvalidStorageId; |
| } |
| |
| return findStorageId(path::normalize(pathInMount)); |
| } |
| |
| IncrementalService::IfsMountPtr IncrementalService::getIfs(StorageId storage) const { |
| std::lock_guard l(mLock); |
| return getIfsLocked(storage); |
| } |
| |
| const IncrementalService::IfsMountPtr& IncrementalService::getIfsLocked(StorageId storage) const { |
| auto it = mMounts.find(storage); |
| if (it == mMounts.end()) { |
| static const base::NoDestructor<IfsMountPtr> kEmpty{}; |
| return *kEmpty; |
| } |
| return it->second; |
| } |
| |
| int IncrementalService::bind(StorageId storage, std::string_view source, std::string_view target, |
| BindKind kind) { |
| if (!isValidMountTarget(target)) { |
| LOG(ERROR) << __func__ << ": not a valid bind target " << target; |
| return -EINVAL; |
| } |
| |
| const auto ifs = getIfs(storage); |
| if (!ifs) { |
| LOG(ERROR) << __func__ << ": no ifs object for storage " << storage; |
| return -EINVAL; |
| } |
| |
| std::unique_lock l(ifs->lock); |
| const auto storageInfo = ifs->storages.find(storage); |
| if (storageInfo == ifs->storages.end()) { |
| LOG(ERROR) << "no storage"; |
| return -EINVAL; |
| } |
| std::string normSource = normalizePathToStorageLocked(*ifs, storageInfo, source); |
| if (normSource.empty()) { |
| LOG(ERROR) << "invalid source path"; |
| return -EINVAL; |
| } |
| l.unlock(); |
| std::unique_lock l2(mLock, std::defer_lock); |
| return addBindMount(*ifs, storage, storageInfo->second.name, std::move(normSource), |
| path::normalize(target), kind, l2); |
| } |
| |
| int IncrementalService::unbind(StorageId storage, std::string_view target) { |
| if (!path::isAbsolute(target)) { |
| return -EINVAL; |
| } |
| |
| LOG(INFO) << "Removing bind point " << target << " for storage " << storage; |
| |
| // Here we should only look up by the exact target, not by a subdirectory of any existing mount, |
| // otherwise there's a chance to unmount something completely unrelated |
| const auto norm = path::normalize(target); |
| std::unique_lock l(mLock); |
| const auto storageIt = mBindsByPath.find(norm); |
| if (storageIt == mBindsByPath.end() || storageIt->second->second.storage != storage) { |
| return -EINVAL; |
| } |
| const auto bindIt = storageIt->second; |
| const auto storageId = bindIt->second.storage; |
| const auto ifs = getIfsLocked(storageId); |
| if (!ifs) { |
| LOG(ERROR) << "Internal error: storageId " << storageId << " for bound path " << target |
| << " is missing"; |
| return -EFAULT; |
| } |
| mBindsByPath.erase(storageIt); |
| l.unlock(); |
| |
| mVold->unmountIncFs(bindIt->first); |
| std::unique_lock l2(ifs->lock); |
| if (ifs->bindPoints.size() <= 1) { |
| ifs->bindPoints.clear(); |
| deleteStorageLocked(*ifs, std::move(l2)); |
| } else { |
| const std::string savedFile = std::move(bindIt->second.savedFilename); |
| ifs->bindPoints.erase(bindIt); |
| l2.unlock(); |
| if (!savedFile.empty()) { |
| mIncFs->unlink(ifs->control, path::join(ifs->root, constants().mount, savedFile)); |
| } |
| } |
| |
| return 0; |
| } |
| |
| std::string IncrementalService::normalizePathToStorageLocked( |
| const IncFsMount& incfs, IncFsMount::StorageMap::const_iterator storageIt, |
| std::string_view path) const { |
| if (!path::isAbsolute(path)) { |
| return path::normalize(path::join(storageIt->second.name, path)); |
| } |
| auto normPath = path::normalize(path); |
| if (path::startsWith(normPath, storageIt->second.name)) { |
| return normPath; |
| } |
| // not that easy: need to find if any of the bind points match |
| const auto bindIt = findParentPath(incfs.bindPoints, normPath); |
| if (bindIt == incfs.bindPoints.end()) { |
| return {}; |
| } |
| return path::join(bindIt->second.sourceDir, path::relativize(bindIt->first, normPath)); |
| } |
| |
| std::string IncrementalService::normalizePathToStorage(const IncFsMount& ifs, StorageId storage, |
| std::string_view path) const { |
| std::unique_lock l(ifs.lock); |
| const auto storageInfo = ifs.storages.find(storage); |
| if (storageInfo == ifs.storages.end()) { |
| return {}; |
| } |
| return normalizePathToStorageLocked(ifs, storageInfo, path); |
| } |
| |
| int IncrementalService::makeFile(StorageId storage, std::string_view path, int mode, FileId id, |
| incfs::NewFileParams params, std::span<const uint8_t> data) { |
| const auto ifs = getIfs(storage); |
| if (!ifs) { |
| return -EINVAL; |
| } |
| if ((IncFsSize)data.size() > params.size) { |
| LOG(ERROR) << "Bad data size - bigger than file size"; |
| return -EINVAL; |
| } |
| if (!data.empty() && (IncFsSize)data.size() != params.size) { |
| // Writing a page is an irreversible operation, and it can't be updated with additional |
| // data later. Check that the last written page is complete, or we may break the file. |
| if (!isPageAligned(data.size())) { |
| LOG(ERROR) << "Bad data size - tried to write half a page?"; |
| return -EINVAL; |
| } |
| } |
| const std::string normPath = normalizePathToStorage(*ifs, storage, path); |
| if (normPath.empty()) { |
| LOG(ERROR) << "Internal error: storageId " << storage << " failed to normalize: " << path; |
| return -EINVAL; |
| } |
| if (auto err = mIncFs->makeFile(ifs->control, normPath, mode, id, params); err) { |
| LOG(ERROR) << "Internal error: storageId " << storage << " failed to makeFile [" << normPath |
| << "]: " << err; |
| return err; |
| } |
| if (params.size > 0) { |
| if (auto err = mIncFs->reserveSpace(ifs->control, id, params.size)) { |
| if (err != -EOPNOTSUPP) { |
| LOG(ERROR) << "Failed to reserve space for a new file: " << err; |
| (void)mIncFs->unlink(ifs->control, normPath); |
| return err; |
| } else { |
| LOG(WARNING) << "Reserving space for backing file isn't supported, " |
| "may run out of disk later"; |
| } |
| } |
| if (!data.empty()) { |
| if (auto err = setFileContent(ifs, id, path, data); err) { |
| (void)mIncFs->unlink(ifs->control, normPath); |
| return err; |
| } |
| } |
| } |
| return 0; |
| } |
| |
| int IncrementalService::makeDir(StorageId storageId, std::string_view path, int mode) { |
| if (auto ifs = getIfs(storageId)) { |
| std::string normPath = normalizePathToStorage(*ifs, storageId, path); |
| if (normPath.empty()) { |
| return -EINVAL; |
| } |
| return mIncFs->makeDir(ifs->control, normPath, mode); |
| } |
| return -EINVAL; |
| } |
| |
| int IncrementalService::makeDirs(StorageId storageId, std::string_view path, int mode) { |
| const auto ifs = getIfs(storageId); |
| if (!ifs) { |
| return -EINVAL; |
| } |
| return makeDirs(*ifs, storageId, path, mode); |
| } |
| |
| int IncrementalService::makeDirs(const IncFsMount& ifs, StorageId storageId, std::string_view path, |
| int mode) { |
| std::string normPath = normalizePathToStorage(ifs, storageId, path); |
| if (normPath.empty()) { |
| return -EINVAL; |
| } |
| return mIncFs->makeDirs(ifs.control, normPath, mode); |
| } |
| |
| int IncrementalService::link(StorageId sourceStorageId, std::string_view oldPath, |
| StorageId destStorageId, std::string_view newPath) { |
| std::unique_lock l(mLock); |
| auto ifsSrc = getIfsLocked(sourceStorageId); |
| if (!ifsSrc) { |
| return -EINVAL; |
| } |
| if (sourceStorageId != destStorageId && getIfsLocked(destStorageId) != ifsSrc) { |
| return -EINVAL; |
| } |
| l.unlock(); |
| std::string normOldPath = normalizePathToStorage(*ifsSrc, sourceStorageId, oldPath); |
| std::string normNewPath = normalizePathToStorage(*ifsSrc, destStorageId, newPath); |
| if (normOldPath.empty() || normNewPath.empty()) { |
| LOG(ERROR) << "Invalid paths in link(): " << normOldPath << " | " << normNewPath; |
| return -EINVAL; |
| } |
| if (auto err = mIncFs->link(ifsSrc->control, normOldPath, normNewPath); err < 0) { |
| PLOG(ERROR) << "Failed to link " << oldPath << "[" << normOldPath << "]" |
| << " to " << newPath << "[" << normNewPath << "]"; |
| return err; |
| } |
| return 0; |
| } |
| |
| int IncrementalService::unlink(StorageId storage, std::string_view path) { |
| if (auto ifs = getIfs(storage)) { |
| std::string normOldPath = normalizePathToStorage(*ifs, storage, path); |
| return mIncFs->unlink(ifs->control, normOldPath); |
| } |
| return -EINVAL; |
| } |
| |
| int IncrementalService::addBindMount(IncFsMount& ifs, StorageId storage, |
| std::string_view storageRoot, std::string&& source, |
| std::string&& target, BindKind kind, |
| std::unique_lock<std::mutex>& mainLock) { |
| if (!isValidMountTarget(target)) { |
| LOG(ERROR) << __func__ << ": invalid mount target " << target; |
| return -EINVAL; |
| } |
| |
| std::string mdFileName; |
| std::string metadataFullPath; |
| if (kind != BindKind::Temporary) { |
| metadata::BindPoint bp; |
| bp.set_storage_id(storage); |
| bp.set_allocated_dest_path(&target); |
| bp.set_allocated_source_subdir(&source); |
| const auto metadata = bp.SerializeAsString(); |
| static_cast<void>(bp.release_dest_path()); |
| static_cast<void>(bp.release_source_subdir()); |
| mdFileName = makeBindMdName(); |
| metadataFullPath = path::join(ifs.root, constants().mount, mdFileName); |
| auto node = mIncFs->makeFile(ifs.control, metadataFullPath, 0444, idFromMetadata(metadata), |
| {.metadata = {metadata.data(), (IncFsSize)metadata.size()}}); |
| if (node) { |
| LOG(ERROR) << __func__ << ": couldn't create a mount node " << mdFileName; |
| return int(node); |
| } |
| } |
| |
| const auto res = addBindMountWithMd(ifs, storage, std::move(mdFileName), std::move(source), |
| std::move(target), kind, mainLock); |
| if (res) { |
| mIncFs->unlink(ifs.control, metadataFullPath); |
| } |
| return res; |
| } |
| |
| int IncrementalService::addBindMountWithMd(IncrementalService::IncFsMount& ifs, StorageId storage, |
| std::string&& metadataName, std::string&& source, |
| std::string&& target, BindKind kind, |
| std::unique_lock<std::mutex>& mainLock) { |
| { |
| std::lock_guard l(mMountOperationLock); |
| const auto status = mVold->bindMount(source, target); |
| if (!status.isOk()) { |
| LOG(ERROR) << "Calling Vold::bindMount() failed: " << status.toString8(); |
| return status.exceptionCode() == binder::Status::EX_SERVICE_SPECIFIC |
| ? status.serviceSpecificErrorCode() > 0 ? -status.serviceSpecificErrorCode() |
| : status.serviceSpecificErrorCode() == 0 |
| ? -EFAULT |
| : status.serviceSpecificErrorCode() |
| : -EIO; |
| } |
| } |
| |
| if (!mainLock.owns_lock()) { |
| mainLock.lock(); |
| } |
| std::lock_guard l(ifs.lock); |
| addBindMountRecordLocked(ifs, storage, std::move(metadataName), std::move(source), |
| std::move(target), kind); |
| return 0; |
| } |
| |
| void IncrementalService::addBindMountRecordLocked(IncFsMount& ifs, StorageId storage, |
| std::string&& metadataName, std::string&& source, |
| std::string&& target, BindKind kind) { |
| const auto [it, _] = |
| ifs.bindPoints.insert_or_assign(target, |
| IncFsMount::Bind{storage, std::move(metadataName), |
| std::move(source), kind}); |
| mBindsByPath[std::move(target)] = it; |
| } |
| |
| RawMetadata IncrementalService::getMetadata(StorageId storage, std::string_view path) const { |
| const auto ifs = getIfs(storage); |
| if (!ifs) { |
| return {}; |
| } |
| const auto normPath = normalizePathToStorage(*ifs, storage, path); |
| if (normPath.empty()) { |
| return {}; |
| } |
| return mIncFs->getMetadata(ifs->control, normPath); |
| } |
| |
| RawMetadata IncrementalService::getMetadata(StorageId storage, FileId node) const { |
| const auto ifs = getIfs(storage); |
| if (!ifs) { |
| return {}; |
| } |
| return mIncFs->getMetadata(ifs->control, node); |
| } |
| |
| void IncrementalService::setUidReadTimeouts(StorageId storage, |
| std::vector<PerUidReadTimeouts>&& perUidReadTimeouts) { |
| using microseconds = std::chrono::microseconds; |
| using milliseconds = std::chrono::milliseconds; |
| |
| auto maxPendingTimeUs = microseconds(0); |
| for (const auto& timeouts : perUidReadTimeouts) { |
| maxPendingTimeUs = std::max(maxPendingTimeUs, microseconds(timeouts.maxPendingTimeUs)); |
| } |
| if (maxPendingTimeUs < Constants::minPerUidTimeout) { |
| LOG(ERROR) << "Skip setting read timeouts (maxPendingTime < Constants::minPerUidTimeout): " |
| << duration_cast<milliseconds>(maxPendingTimeUs).count() << "ms < " |
| << Constants::minPerUidTimeout.count() << "ms"; |
| return; |
| } |
| |
| const auto ifs = getIfs(storage); |
| if (!ifs) { |
| LOG(ERROR) << "Setting read timeouts failed: invalid storage id: " << storage; |
| return; |
| } |
| |
| if (auto err = mIncFs->setUidReadTimeouts(ifs->control, perUidReadTimeouts); err < 0) { |
| LOG(ERROR) << "Setting read timeouts failed: " << -err; |
| return; |
| } |
| |
| const auto timeout = Clock::now() + maxPendingTimeUs - Constants::perUidTimeoutOffset; |
| addIfsStateCallback(storage, [this, timeout](StorageId storageId, IfsState state) -> bool { |
| if (checkUidReadTimeouts(storageId, state, timeout)) { |
| return true; |
| } |
| clearUidReadTimeouts(storageId); |
| return false; |
| }); |
| } |
| |
| void IncrementalService::clearUidReadTimeouts(StorageId storage) { |
| const auto ifs = getIfs(storage); |
| if (!ifs) { |
| return; |
| } |
| mIncFs->setUidReadTimeouts(ifs->control, {}); |
| } |
| |
| bool IncrementalService::checkUidReadTimeouts(StorageId storage, IfsState state, |
| Clock::time_point timeLimit) { |
| if (Clock::now() >= timeLimit) { |
| // Reached maximum timeout. |
| return false; |
| } |
| if (state.error) { |
| // Something is wrong, abort. |
| return false; |
| } |
| |
| // Still loading? |
| if (state.fullyLoaded && !state.readLogsEnabled) { |
| return false; |
| } |
| |
| const auto timeLeft = timeLimit - Clock::now(); |
| if (timeLeft < Constants::progressUpdateInterval) { |
| // Don't bother. |
| return false; |
| } |
| |
| return true; |
| } |
| |
| std::unordered_set<std::string_view> IncrementalService::adoptMountedInstances() { |
| std::unordered_set<std::string_view> mountedRootNames; |
| mIncFs->listExistingMounts([this, &mountedRootNames](auto root, auto backingDir, auto binds) { |
| LOG(INFO) << "Existing mount: " << backingDir << "->" << root; |
| for (auto [source, target] : binds) { |
| LOG(INFO) << " bind: '" << source << "'->'" << target << "'"; |
| LOG(INFO) << " " << path::join(root, source); |
| } |
| |
| // Ensure it's a kind of a mount that's managed by IncrementalService |
| if (path::basename(root) != constants().mount || |
| path::basename(backingDir) != constants().backing) { |
| return; |
| } |
| const auto expectedRoot = path::dirname(root); |
| if (path::dirname(backingDir) != expectedRoot) { |
| return; |
| } |
| if (path::dirname(expectedRoot) != mIncrementalDir) { |
| return; |
| } |
| if (!path::basename(expectedRoot).starts_with(constants().mountKeyPrefix)) { |
| return; |
| } |
| |
| LOG(INFO) << "Looks like an IncrementalService-owned: " << expectedRoot; |
| |
| // make sure we clean up the mount if it happens to be a bad one. |
| // Note: unmounting needs to run first, so the cleanup object is created _last_. |
| auto cleanupFiles = makeCleanup([&]() { |
| LOG(INFO) << "Failed to adopt existing mount, deleting files: " << expectedRoot; |
| IncFsMount::cleanupFilesystem(expectedRoot); |
| }); |
| auto cleanupMounts = makeCleanup([&]() { |
| LOG(INFO) << "Failed to adopt existing mount, cleaning up: " << expectedRoot; |
| for (auto&& [_, target] : binds) { |
| mVold->unmountIncFs(std::string(target)); |
| } |
| mVold->unmountIncFs(std::string(root)); |
| }); |
| |
| auto control = mIncFs->openMount(root); |
| if (!control) { |
| LOG(INFO) << "failed to open mount " << root; |
| return; |
| } |
| |
| auto mountRecord = |
| parseFromIncfs<metadata::Mount>(mIncFs.get(), control, |
| path::join(root, constants().infoMdName)); |
| if (!mountRecord.has_loader() || !mountRecord.has_storage()) { |
| LOG(ERROR) << "Bad mount metadata in mount at " << expectedRoot; |
| return; |
| } |
| |
| auto mountId = mountRecord.storage().id(); |
| mNextId = std::max(mNextId, mountId + 1); |
| |
| DataLoaderParamsParcel dataLoaderParams; |
| { |
| const auto& loader = mountRecord.loader(); |
| dataLoaderParams.type = (content::pm::DataLoaderType)loader.type(); |
| dataLoaderParams.packageName = loader.package_name(); |
| dataLoaderParams.className = loader.class_name(); |
| dataLoaderParams.arguments = loader.arguments(); |
| } |
| |
| // Not way to obtain a real sysfs key at this point - metrics will stop working after "soft" |
| // reboot. |
| std::string metricsKey{}; |
| auto ifs = std::make_shared<IncFsMount>(std::string(expectedRoot), std::move(metricsKey), |
| mountId, std::move(control), *this); |
| (void)cleanupFiles.release(); // ifs will take care of that now |
| |
| // Check if marker file present. |
| if (checkReadLogsDisabledMarker(root)) { |
| ifs->disallowReadLogs(); |
| } |
| |
| std::vector<std::pair<std::string, metadata::BindPoint>> permanentBindPoints; |
| auto d = openDir(root); |
| while (auto e = ::readdir(d.get())) { |
| if (e->d_type == DT_REG) { |
| auto name = std::string_view(e->d_name); |
| if (name.starts_with(constants().mountpointMdPrefix)) { |
| permanentBindPoints |
| .emplace_back(name, |
| parseFromIncfs<metadata::BindPoint>(mIncFs.get(), |
| ifs->control, |
| path::join(root, |
| name))); |
| if (permanentBindPoints.back().second.dest_path().empty() || |
| permanentBindPoints.back().second.source_subdir().empty()) { |
| permanentBindPoints.pop_back(); |
| mIncFs->unlink(ifs->control, path::join(root, name)); |
| } else { |
| LOG(INFO) << "Permanent bind record: '" |
| << permanentBindPoints.back().second.source_subdir() << "'->'" |
| << permanentBindPoints.back().second.dest_path() << "'"; |
| } |
| } |
| } else if (e->d_type == DT_DIR) { |
| if (e->d_name == "."sv || e->d_name == ".."sv) { |
| continue; |
| } |
| auto name = std::string_view(e->d_name); |
| if (name.starts_with(constants().storagePrefix)) { |
| int storageId; |
| const auto res = |
| std::from_chars(name.data() + constants().storagePrefix.size() + 1, |
| name.data() + name.size(), storageId); |
| if (res.ec != std::errc{} || *res.ptr != '_') { |
| LOG(WARNING) << "Ignoring storage with invalid name '" << name |
| << "' for mount " << expectedRoot; |
| continue; |
| } |
| auto [_, inserted] = mMounts.try_emplace(storageId, ifs); |
| if (!inserted) { |
| LOG(WARNING) << "Ignoring storage with duplicate id " << storageId |
| << " for mount " << expectedRoot; |
| continue; |
| } |
| ifs->storages.insert_or_assign(storageId, |
| IncFsMount::Storage{path::join(root, name)}); |
| mNextId = std::max(mNextId, storageId + 1); |
| } |
| } |
| } |
| |
| if (ifs->storages.empty()) { |
| LOG(WARNING) << "No valid storages in mount " << root; |
| return; |
| } |
| |
| // now match the mounted directories with what we expect to have in the metadata |
| { |
| std::unique_lock l(mLock, std::defer_lock); |
| for (auto&& [metadataFile, bindRecord] : permanentBindPoints) { |
| auto mountedIt = std::find_if(binds.begin(), binds.end(), |
| [&, bindRecord = bindRecord](auto&& bind) { |
| return bind.second == bindRecord.dest_path() && |
| path::join(root, bind.first) == |
| bindRecord.source_subdir(); |
| }); |
| if (mountedIt != binds.end()) { |
| LOG(INFO) << "Matched permanent bound " << bindRecord.source_subdir() |
| << " to mount " << mountedIt->first; |
| addBindMountRecordLocked(*ifs, bindRecord.storage_id(), std::move(metadataFile), |
| std::move(*bindRecord.mutable_source_subdir()), |
| std::move(*bindRecord.mutable_dest_path()), |
| BindKind::Permanent); |
| if (mountedIt != binds.end() - 1) { |
| std::iter_swap(mountedIt, binds.end() - 1); |
| } |
| binds = binds.first(binds.size() - 1); |
| } else { |
| LOG(INFO) << "Didn't match permanent bound " << bindRecord.source_subdir() |
| << ", mounting"; |
| // doesn't exist - try mounting back |
| if (addBindMountWithMd(*ifs, bindRecord.storage_id(), std::move(metadataFile), |
| std::move(*bindRecord.mutable_source_subdir()), |
| std::move(*bindRecord.mutable_dest_path()), |
| BindKind::Permanent, l)) { |
| mIncFs->unlink(ifs->control, metadataFile); |
| } |
| } |
| } |
| } |
| |
| // if anything stays in |binds| those are probably temporary binds; system restarted since |
| // they were mounted - so let's unmount them all. |
| for (auto&& [source, target] : binds) { |
| if (source.empty()) { |
| continue; |
| } |
| mVold->unmountIncFs(std::string(target)); |
| } |
| (void)cleanupMounts.release(); // ifs now manages everything |
| |
| if (ifs->bindPoints.empty()) { |
| LOG(WARNING) << "No valid bind points for mount " << expectedRoot; |
| deleteStorage(*ifs); |
| return; |
| } |
| |
| prepareDataLoaderLocked(*ifs, std::move(dataLoaderParams)); |
| CHECK(ifs->dataLoaderStub); |
| |
| mountedRootNames.insert(path::basename(ifs->root)); |
| |
| // not locking here at all: we're still in the constructor, no other calls can happen |
| mMounts[ifs->mountId] = std::move(ifs); |
| }); |
| |
| return mountedRootNames; |
| } |
| |
| void IncrementalService::mountExistingImages( |
| const std::unordered_set<std::string_view>& mountedRootNames) { |
| auto dir = openDir(mIncrementalDir); |
| if (!dir) { |
| PLOG(WARNING) << "Couldn't open the root incremental dir " << mIncrementalDir; |
| return; |
| } |
| while (auto entry = ::readdir(dir.get())) { |
| if (entry->d_type != DT_DIR) { |
| continue; |
| } |
| std::string_view name = entry->d_name; |
| if (!name.starts_with(constants().mountKeyPrefix)) { |
| continue; |
| } |
| if (mountedRootNames.find(name) != mountedRootNames.end()) { |
| continue; |
| } |
| const auto root = path::join(mIncrementalDir, name); |
| if (!mountExistingImage(root)) { |
| IncFsMount::cleanupFilesystem(root); |
| } |
| } |
| } |
| |
| bool IncrementalService::mountExistingImage(std::string_view root) { |
| auto mountTarget = path::join(root, constants().mount); |
| const auto backing = path::join(root, constants().backing); |
| std::string mountKey(path::basename(path::dirname(mountTarget))); |
| |
| IncrementalFileSystemControlParcel controlParcel; |
| auto metricsKey = makeUniqueName(mountKey); |
| auto status = mVold->mountIncFs(backing, mountTarget, 0, metricsKey, &controlParcel); |
| if (!status.isOk()) { |
| LOG(ERROR) << "Vold::mountIncFs() failed: " << status.toString8(); |
| return false; |
| } |
| |
| int cmd = controlParcel.cmd.release().release(); |
| int pendingReads = controlParcel.pendingReads.release().release(); |
| int logs = controlParcel.log.release().release(); |
| int blocksWritten = |
| controlParcel.blocksWritten ? controlParcel.blocksWritten->release().release() : -1; |
| IncFsMount::Control control = mIncFs->createControl(cmd, pendingReads, logs, blocksWritten); |
| |
| auto ifs = std::make_shared<IncFsMount>(std::string(root), std::move(metricsKey), -1, |
| std::move(control), *this); |
| |
| auto mount = parseFromIncfs<metadata::Mount>(mIncFs.get(), ifs->control, |
| path::join(mountTarget, constants().infoMdName)); |
| if (!mount.has_loader() || !mount.has_storage()) { |
| LOG(ERROR) << "Bad mount metadata in mount at " << root; |
| return false; |
| } |
| |
| ifs->mountId = mount.storage().id(); |
| mNextId = std::max(mNextId, ifs->mountId + 1); |
| |
| // Check if marker file present. |
| if (checkReadLogsDisabledMarker(mountTarget)) { |
| ifs->disallowReadLogs(); |
| } |
| |
| // DataLoader params |
| DataLoaderParamsParcel dataLoaderParams; |
| { |
| const auto& loader = mount.loader(); |
| dataLoaderParams.type = (content::pm::DataLoaderType)loader.type(); |
| dataLoaderParams.packageName = loader.package_name(); |
| dataLoaderParams.className = loader.class_name(); |
| dataLoaderParams.arguments = loader.arguments(); |
| } |
| |
| prepareDataLoaderLocked(*ifs, std::move(dataLoaderParams)); |
| CHECK(ifs->dataLoaderStub); |
| |
| std::vector<std::pair<std::string, metadata::BindPoint>> bindPoints; |
| auto d = openDir(mountTarget); |
| while (auto e = ::readdir(d.get())) { |
| if (e->d_type == DT_REG) { |
| auto name = std::string_view(e->d_name); |
| if (name.starts_with(constants().mountpointMdPrefix)) { |
| bindPoints.emplace_back(name, |
| parseFromIncfs<metadata::BindPoint>(mIncFs.get(), |
| ifs->control, |
| path::join(mountTarget, |
| name))); |
| if (bindPoints.back().second.dest_path().empty() || |
| bindPoints.back().second.source_subdir().empty()) { |
| bindPoints.pop_back(); |
| mIncFs->unlink(ifs->control, path::join(ifs->root, constants().mount, name)); |
| } |
| } |
| } else if (e->d_type == DT_DIR) { |
| if (e->d_name == "."sv || e->d_name == ".."sv) { |
| continue; |
| } |
| auto name = std::string_view(e->d_name); |
| if (name.starts_with(constants().storagePrefix)) { |
| int storageId; |
| const auto res = std::from_chars(name.data() + constants().storagePrefix.size() + 1, |
| name.data() + name.size(), storageId); |
| if (res.ec != std::errc{} || *res.ptr != '_') { |
| LOG(WARNING) << "Ignoring storage with invalid name '" << name << "' for mount " |
| << root; |
| continue; |
| } |
| auto [_, inserted] = mMounts.try_emplace(storageId, ifs); |
| if (!inserted) { |
| LOG(WARNING) << "Ignoring storage with duplicate id " << storageId |
| << " for mount " << root; |
| continue; |
| } |
| ifs->storages.insert_or_assign(storageId, |
| IncFsMount::Storage{ |
| path::join(root, constants().mount, name)}); |
| mNextId = std::max(mNextId, storageId + 1); |
| } |
| } |
| } |
| |
| if (ifs->storages.empty()) { |
| LOG(WARNING) << "No valid storages in mount " << root; |
| return false; |
| } |
| |
| int bindCount = 0; |
| { |
| std::unique_lock l(mLock, std::defer_lock); |
| for (auto&& bp : bindPoints) { |
| bindCount += !addBindMountWithMd(*ifs, bp.second.storage_id(), std::move(bp.first), |
| std::move(*bp.second.mutable_source_subdir()), |
| std::move(*bp.second.mutable_dest_path()), |
| BindKind::Permanent, l); |
| } |
| } |
| |
| if (bindCount == 0) { |
| LOG(WARNING) << "No valid bind points for mount " << root; |
| deleteStorage(*ifs); |
| return false; |
| } |
| |
| // not locking here at all: we're still in the constructor, no other calls can happen |
| mMounts[ifs->mountId] = std::move(ifs); |
| return true; |
| } |
| |
| void IncrementalService::runCmdLooper() { |
| constexpr auto kTimeoutMsecs = -1; |
| while (mRunning.load(std::memory_order_relaxed)) { |
| mLooper->pollAll(kTimeoutMsecs); |
| } |
| } |
| |
| void IncrementalService::trimReservedSpaceV1(const IncFsMount& ifs) { |
| mIncFs->forEachFile(ifs.control, [this](auto&& control, auto&& fileId) { |
| if (mIncFs->isFileFullyLoaded(control, fileId) == incfs::LoadingState::Full) { |
| mIncFs->reserveSpace(control, fileId, -1); |
| } |
| return true; |
| }); |
| } |
| |
| void IncrementalService::prepareDataLoaderLocked(IncFsMount& ifs, DataLoaderParamsParcel&& params, |
| DataLoaderStatusListener&& statusListener, |
| const StorageHealthCheckParams& healthCheckParams, |
| StorageHealthListener&& healthListener) { |
| FileSystemControlParcel fsControlParcel; |
| fsControlParcel.incremental = std::make_optional<IncrementalFileSystemControlParcel>(); |
| fsControlParcel.incremental->cmd.reset(dup(ifs.control.cmd())); |
| fsControlParcel.incremental->pendingReads.reset(dup(ifs.control.pendingReads())); |
| fsControlParcel.incremental->log.reset(dup(ifs.control.logs())); |
| if (ifs.control.blocksWritten() >= 0) { |
| fsControlParcel.incremental->blocksWritten.emplace(dup(ifs.control.blocksWritten())); |
| } |
| fsControlParcel.service = new IncrementalServiceConnector(*this, ifs.mountId); |
| |
| ifs.dataLoaderStub = |
| new DataLoaderStub(*this, ifs.mountId, std::move(params), std::move(fsControlParcel), |
| std::move(statusListener), healthCheckParams, |
| std::move(healthListener), path::join(ifs.root, constants().mount)); |
| |
| // pre-v2 IncFS doesn't do automatic reserved space trimming - need to run it manually |
| if (!(mIncFs->features() & incfs::Features::v2)) { |
| addIfsStateCallback(ifs.mountId, [this](StorageId storageId, IfsState state) -> bool { |
| if (!state.fullyLoaded) { |
| return true; |
| } |
| |
| const auto ifs = getIfs(storageId); |
| if (!ifs) { |
| return false; |
| } |
| trimReservedSpaceV1(*ifs); |
| return false; |
| }); |
| } |
| |
| addIfsStateCallback(ifs.mountId, [this](StorageId storageId, IfsState state) -> bool { |
| if (!state.fullyLoaded || state.readLogsEnabled) { |
| return true; |
| } |
| |
| DataLoaderStubPtr dataLoaderStub; |
| { |
| const auto ifs = getIfs(storageId); |
| if (!ifs) { |
| return false; |
| } |
| |
| std::unique_lock l(ifs->lock); |
| dataLoaderStub = std::exchange(ifs->dataLoaderStub, nullptr); |
| } |
| |
| if (dataLoaderStub) { |
| dataLoaderStub->cleanupResources(); |
| } |
| |
| return false; |
| }); |
| } |
| |
| template <class Duration> |
| static constexpr auto castToMs(Duration d) { |
| return std::chrono::duration_cast<std::chrono::milliseconds>(d); |
| } |
| |
| // Extract lib files from zip, create new files in incfs and write data to them |
| // Lib files should be placed next to the APK file in the following matter: |
| // Example: |
| // /path/to/base.apk |
| // /path/to/lib/arm/first.so |
| // /path/to/lib/arm/second.so |
| bool IncrementalService::configureNativeBinaries(StorageId storage, std::string_view apkFullPath, |
| std::string_view libDirRelativePath, |
| std::string_view abi, bool extractNativeLibs) { |
| auto start = Clock::now(); |
| |
| const auto ifs = getIfs(storage); |
| if (!ifs) { |
| LOG(ERROR) << "Invalid storage " << storage; |
| return false; |
| } |
| |
| const auto targetLibPathRelativeToStorage = |
| path::join(path::dirname(normalizePathToStorage(*ifs, storage, apkFullPath)), |
| libDirRelativePath); |
| |
| // First prepare target directories if they don't exist yet |
| if (auto res = makeDirs(*ifs, storage, targetLibPathRelativeToStorage, 0755)) { |
| LOG(ERROR) << "Failed to prepare target lib directory " << targetLibPathRelativeToStorage |
| << " errno: " << res; |
| return false; |
| } |
| |
| auto mkDirsTs = Clock::now(); |
| ZipArchiveHandle zipFileHandle; |
| if (OpenArchive(path::c_str(apkFullPath), &zipFileHandle)) { |
| LOG(ERROR) << "Failed to open zip file at " << apkFullPath; |
| return false; |
| } |
| |
| // Need a shared pointer: will be passing it into all unpacking jobs. |
| std::shared_ptr<ZipArchive> zipFile(zipFileHandle, [](ZipArchiveHandle h) { CloseArchive(h); }); |
| void* cookie = nullptr; |
| const auto libFilePrefix = path::join(constants().libDir, abi) += "/"; |
| if (StartIteration(zipFile.get(), &cookie, libFilePrefix, constants().libSuffix)) { |
| LOG(ERROR) << "Failed to start zip iteration for " << apkFullPath; |
| return false; |
| } |
| auto endIteration = [](void* cookie) { EndIteration(cookie); }; |
| auto iterationCleaner = std::unique_ptr<void, decltype(endIteration)>(cookie, endIteration); |
| |
| auto openZipTs = Clock::now(); |
| |
| auto mapFiles = (mIncFs->features() & incfs::Features::v2); |
| incfs::FileId sourceId; |
| if (mapFiles) { |
| sourceId = mIncFs->getFileId(ifs->control, apkFullPath); |
| if (!incfs::isValidFileId(sourceId)) { |
| LOG(WARNING) << "Error getting IncFS file ID for apk path '" << apkFullPath |
| << "', mapping disabled"; |
| mapFiles = false; |
| } |
| } |
| |
| std::vector<Job> jobQueue; |
| ZipEntry entry; |
| std::string_view fileName; |
| while (!Next(cookie, &entry, &fileName)) { |
| if (fileName.empty()) { |
| continue; |
| } |
| |
| const auto entryUncompressed = entry.method == kCompressStored; |
| const auto entryPageAligned = isPageAligned(entry.offset); |
| |
| if (!extractNativeLibs) { |
| // ensure the file is properly aligned and unpacked |
| if (!entryUncompressed) { |
| LOG(WARNING) << "Library " << fileName << " must be uncompressed to mmap it"; |
| return false; |
| } |
| if (!entryPageAligned) { |
| LOG(WARNING) << "Library " << fileName |
| << " must be page-aligned to mmap it, offset = 0x" << std::hex |
| << entry.offset; |
| return false; |
| } |
| continue; |
| } |
| |
| auto startFileTs = Clock::now(); |
| |
| const auto libName = path::basename(fileName); |
| auto targetLibPath = path::join(targetLibPathRelativeToStorage, libName); |
| const auto targetLibPathAbsolute = normalizePathToStorage(*ifs, storage, targetLibPath); |
| // If the extract file already exists, skip |
| if (access(targetLibPathAbsolute.c_str(), F_OK) == 0) { |
| if (perfLoggingEnabled()) { |
| LOG(INFO) << "incfs: Native lib file already exists: " << targetLibPath |
| << "; skipping extraction, spent " |
| << elapsedMcs(startFileTs, Clock::now()) << "mcs"; |
| } |
| continue; |
| } |
| |
| if (mapFiles && entryUncompressed && entryPageAligned && entry.uncompressed_length > 0) { |
| incfs::NewMappedFileParams mappedFileParams = { |
| .sourceId = sourceId, |
| .sourceOffset = entry.offset, |
| .size = entry.uncompressed_length, |
| }; |
| |
| if (auto res = mIncFs->makeMappedFile(ifs->control, targetLibPathAbsolute, 0755, |
| mappedFileParams); |
| res == 0) { |
| if (perfLoggingEnabled()) { |
| auto doneTs = Clock::now(); |
| LOG(INFO) << "incfs: Mapped " << libName << ": " |
| << elapsedMcs(startFileTs, doneTs) << "mcs"; |
| } |
| continue; |
| } else { |
| LOG(WARNING) << "Failed to map file for: '" << targetLibPath << "' errno: " << res |
| << "; falling back to full extraction"; |
| } |
| } |
| |
| // Create new lib file without signature info |
| incfs::NewFileParams libFileParams = { |
| .size = entry.uncompressed_length, |
| .signature = {}, |
| // Metadata of the new lib file is its relative path |
| .metadata = {targetLibPath.c_str(), (IncFsSize)targetLibPath.size()}, |
| }; |
| incfs::FileId libFileId = idFromMetadata(targetLibPath); |
| if (auto res = mIncFs->makeFile(ifs->control, targetLibPathAbsolute, 0755, libFileId, |
| libFileParams)) { |
| LOG(ERROR) << "Failed to make file for: " << targetLibPath << " errno: " << res; |
| // If one lib file fails to be created, abort others as well |
| return false; |
| } |
| |
| auto makeFileTs = Clock::now(); |
| |
| // If it is a zero-byte file, skip data writing |
| if (entry.uncompressed_length == 0) { |
| if (perfLoggingEnabled()) { |
| LOG(INFO) << "incfs: Extracted " << libName |
| << "(0 bytes): " << elapsedMcs(startFileTs, makeFileTs) << "mcs"; |
| } |
| continue; |
| } |
| |
| jobQueue.emplace_back([this, zipFile, entry, ifs = std::weak_ptr<IncFsMount>(ifs), |
| libFileId, libPath = std::move(targetLibPath), |
| makeFileTs]() mutable { |
| extractZipFile(ifs.lock(), zipFile.get(), entry, libFileId, libPath, makeFileTs); |
| }); |
| |
| if (perfLoggingEnabled()) { |
| auto prepareJobTs = Clock::now(); |
| LOG(INFO) << "incfs: Processed " << libName << ": " |
| << elapsedMcs(startFileTs, prepareJobTs) |
| << "mcs, make file: " << elapsedMcs(startFileTs, makeFileTs) |
| << " prepare job: " << elapsedMcs(makeFileTs, prepareJobTs); |
| } |
| } |
| |
| auto processedTs = Clock::now(); |
| |
| if (!jobQueue.empty()) { |
| { |
| std::lock_guard lock(mJobMutex); |
| if (mRunning) { |
| auto& existingJobs = mJobQueue[ifs->mountId]; |
| if (existingJobs.empty()) { |
| existingJobs = std::move(jobQueue); |
| } else { |
| existingJobs.insert(existingJobs.end(), std::move_iterator(jobQueue.begin()), |
| std::move_iterator(jobQueue.end())); |
| } |
| } |
| } |
| mJobCondition.notify_all(); |
| } |
| |
| if (perfLoggingEnabled()) { |
| auto end = Clock::now(); |
| LOG(INFO) << "incfs: configureNativeBinaries complete in " << elapsedMcs(start, end) |
| << "mcs, make dirs: " << elapsedMcs(start, mkDirsTs) |
| << " open zip: " << elapsedMcs(mkDirsTs, openZipTs) |
| << " make files: " << elapsedMcs(openZipTs, processedTs) |
| << " schedule jobs: " << elapsedMcs(processedTs, end); |
| } |
| |
| return true; |
| } |
| |
| void IncrementalService::extractZipFile(const IfsMountPtr& ifs, ZipArchiveHandle zipFile, |
| ZipEntry& entry, const incfs::FileId& libFileId, |
| std::string_view debugLibPath, |
| Clock::time_point scheduledTs) { |
| if (!ifs) { |
| LOG(INFO) << "Skipping zip file " << debugLibPath << " extraction for an expired mount"; |
| return; |
| } |
| |
| auto startedTs = Clock::now(); |
| |
| // Write extracted data to new file |
| // NOTE: don't zero-initialize memory, it may take a while for nothing |
| auto libData = std::unique_ptr<uint8_t[]>(new uint8_t[entry.uncompressed_length]); |
| if (ExtractToMemory(zipFile, &entry, libData.get(), entry.uncompressed_length)) { |
| LOG(ERROR) << "Failed to extract native lib zip entry: " << path::basename(debugLibPath); |
| return; |
| } |
| |
| auto extractFileTs = Clock::now(); |
| |
| if (setFileContent(ifs, libFileId, debugLibPath, |
| std::span(libData.get(), entry.uncompressed_length))) { |
| return; |
| } |
| |
| if (perfLoggingEnabled()) { |
| auto endFileTs = Clock::now(); |
| LOG(INFO) << "incfs: Extracted " << path::basename(debugLibPath) << "(" |
| << entry.compressed_length << " -> " << entry.uncompressed_length |
| << " bytes): " << elapsedMcs(startedTs, endFileTs) |
| << "mcs, scheduling delay: " << elapsedMcs(scheduledTs, startedTs) |
| << " extract: " << elapsedMcs(startedTs, extractFileTs) |
| << " open/prepare/write: " << elapsedMcs(extractFileTs, endFileTs); |
| } |
| } |
| |
| bool IncrementalService::waitForNativeBinariesExtraction(StorageId storage) { |
| struct WaitPrinter { |
| const Clock::time_point startTs = Clock::now(); |
| ~WaitPrinter() noexcept { |
| if (perfLoggingEnabled()) { |
| const auto endTs = Clock::now(); |
| LOG(INFO) << "incfs: waitForNativeBinariesExtraction() complete in " |
| << elapsedMcs(startTs, endTs) << "mcs"; |
| } |
| } |
| } waitPrinter; |
| |
| MountId mount; |
| { |
| auto ifs = getIfs(storage); |
| if (!ifs) { |
| return true; |
| } |
| mount = ifs->mountId; |
| } |
| |
| std::unique_lock lock(mJobMutex); |
| mJobCondition.wait(lock, [this, mount] { |
| return !mRunning || |
| (mPendingJobsMount != mount && mJobQueue.find(mount) == mJobQueue.end()); |
| }); |
| return mRunning; |
| } |
| |
| int IncrementalService::setFileContent(const IfsMountPtr& ifs, const incfs::FileId& fileId, |
| std::string_view debugFilePath, |
| std::span<const uint8_t> data) const { |
| auto startTs = Clock::now(); |
| |
| const auto writeFd = mIncFs->openForSpecialOps(ifs->control, fileId); |
| if (!writeFd.ok()) { |
| LOG(ERROR) << "Failed to open write fd for: " << debugFilePath |
| << " errno: " << writeFd.get(); |
| return writeFd.get(); |
| } |
| |
| const auto dataLength = data.size(); |
| |
| auto openFileTs = Clock::now(); |
| const int numBlocks = (data.size() + constants().blockSize - 1) / constants().blockSize; |
| std::vector<IncFsDataBlock> instructions(numBlocks); |
| for (int i = 0; i < numBlocks; i++) { |
| const auto blockSize = std::min<long>(constants().blockSize, data.size()); |
| instructions[i] = IncFsDataBlock{ |
| .fileFd = writeFd.get(), |
| .pageIndex = static_cast<IncFsBlockIndex>(i), |
| .compression = INCFS_COMPRESSION_KIND_NONE, |
| .kind = INCFS_BLOCK_KIND_DATA, |
| .dataSize = static_cast<uint32_t>(blockSize), |
| .data = reinterpret_cast<const char*>(data.data()), |
| }; |
| data = data.subspan(blockSize); |
| } |
| auto prepareInstsTs = Clock::now(); |
| |
| size_t res = mIncFs->writeBlocks(instructions); |
| if (res != instructions.size()) { |
| LOG(ERROR) << "Failed to write data into: " << debugFilePath; |
| return res; |
| } |
| |
| if (perfLoggingEnabled()) { |
| auto endTs = Clock::now(); |
| LOG(INFO) << "incfs: Set file content " << debugFilePath << "(" << dataLength |
| << " bytes): " << elapsedMcs(startTs, endTs) |
| << "mcs, open: " << elapsedMcs(startTs, openFileTs) |
| << " prepare: " << elapsedMcs(openFileTs, prepareInstsTs) |
| << " write: " << elapsedMcs(prepareInstsTs, endTs); |
| } |
| |
| return 0; |
| } |
| |
| incfs::LoadingState IncrementalService::isFileFullyLoaded(StorageId storage, |
| std::string_view filePath) const { |
| std::unique_lock l(mLock); |
| const auto ifs = getIfsLocked(storage); |
| if (!ifs) { |
| LOG(ERROR) << "isFileFullyLoaded failed, invalid storageId: " << storage; |
| return incfs::LoadingState(-EINVAL); |
| } |
| const auto storageInfo = ifs->storages.find(storage); |
| if (storageInfo == ifs->storages.end()) { |
| LOG(ERROR) << "isFileFullyLoaded failed, no storage: " << storage; |
| return incfs::LoadingState(-EINVAL); |
| } |
| l.unlock(); |
| return mIncFs->isFileFullyLoaded(ifs->control, filePath); |
| } |
| |
| incfs::LoadingState IncrementalService::isMountFullyLoaded(StorageId storage) const { |
| const auto ifs = getIfs(storage); |
| if (!ifs) { |
| LOG(ERROR) << "isMountFullyLoaded failed, invalid storageId: " << storage; |
| return incfs::LoadingState(-EINVAL); |
| } |
| return mIncFs->isEverythingFullyLoaded(ifs->control); |
| } |
| |
| IncrementalService::LoadingProgress IncrementalService::getLoadingProgress( |
| StorageId storage) const { |
| std::unique_lock l(mLock); |
| const auto ifs = getIfsLocked(storage); |
| if (!ifs) { |
| LOG(ERROR) << "getLoadingProgress failed, invalid storageId: " << storage; |
| return {-EINVAL, -EINVAL}; |
| } |
| const auto storageInfo = ifs->storages.find(storage); |
| if (storageInfo == ifs->storages.end()) { |
| LOG(ERROR) << "getLoadingProgress failed, no storage: " << storage; |
| return {-EINVAL, -EINVAL}; |
| } |
| l.unlock(); |
| return getLoadingProgressFromPath(*ifs, storageInfo->second.name); |
| } |
| |
| IncrementalService::LoadingProgress IncrementalService::getLoadingProgressFromPath( |
| const IncFsMount& ifs, std::string_view storagePath) const { |
| ssize_t totalBlocks = 0, filledBlocks = 0, error = 0; |
| mFs->listFilesRecursive(storagePath, [&, this](auto filePath) { |
| const auto [filledBlocksCount, totalBlocksCount] = |
| mIncFs->countFilledBlocks(ifs.control, filePath); |
| if (filledBlocksCount == -EOPNOTSUPP || filledBlocksCount == -ENOTSUP || |
| filledBlocksCount == -ENOENT) { |
| // a kind of a file that's not really being loaded, e.g. a mapped range |
| // an older IncFS used to return ENOENT in this case, so handle it the same way |
| return true; |
| } |
| if (filledBlocksCount < 0) { |
| LOG(ERROR) << "getLoadingProgress failed to get filled blocks count for: " << filePath |
| << ", errno: " << filledBlocksCount; |
| error = filledBlocksCount; |
| return false; |
| } |
| totalBlocks += totalBlocksCount; |
| filledBlocks += filledBlocksCount; |
| return true; |
| }); |
| |
| return error ? LoadingProgress{error, error} : LoadingProgress{filledBlocks, totalBlocks}; |
| } |
| |
| bool IncrementalService::updateLoadingProgress(StorageId storage, |
| StorageLoadingProgressListener&& progressListener) { |
| const auto progress = getLoadingProgress(storage); |
| if (progress.isError()) { |
| // Failed to get progress from incfs, abort. |
| return false; |
| } |
| progressListener->onStorageLoadingProgressChanged(storage, progress.getProgress()); |
| if (progress.fullyLoaded()) { |
| // Stop updating progress once it is fully loaded |
| return true; |
| } |
| addTimedJob(*mProgressUpdateJobQueue, storage, |
| Constants::progressUpdateInterval /* repeat after 1s */, |
| [storage, progressListener = std::move(progressListener), this]() mutable { |
| updateLoadingProgress(storage, std::move(progressListener)); |
| }); |
| return true; |
| } |
| |
| bool IncrementalService::registerLoadingProgressListener( |
| StorageId storage, StorageLoadingProgressListener progressListener) { |
| return updateLoadingProgress(storage, std::move(progressListener)); |
| } |
| |
| bool IncrementalService::unregisterLoadingProgressListener(StorageId storage) { |
| return removeTimedJobs(*mProgressUpdateJobQueue, storage); |
| } |
| |
| bool IncrementalService::perfLoggingEnabled() { |
| static const bool enabled = base::GetBoolProperty("incremental.perflogging", false); |
| return enabled; |
| } |
| |
| void IncrementalService::runJobProcessing() { |
| for (;;) { |
| std::unique_lock lock(mJobMutex); |
| mJobCondition.wait(lock, [this]() { return !mRunning || !mJobQueue.empty(); }); |
| if (!mRunning) { |
| return; |
| } |
| |
| auto it = mJobQueue.begin(); |
| mPendingJobsMount = it->first; |
| auto queue = std::move(it->second); |
| mJobQueue.erase(it); |
| lock.unlock(); |
| |
| for (auto&& job : queue) { |
| job(); |
| } |
| |
| lock.lock(); |
| mPendingJobsMount = kInvalidStorageId; |
| lock.unlock(); |
| mJobCondition.notify_all(); |
| } |
| } |
| |
| void IncrementalService::registerAppOpsCallback(const std::string& packageName) { |
| sp<IAppOpsCallback> listener; |
| { |
| std::unique_lock lock{mCallbacksLock}; |
| auto& cb = mCallbackRegistered[packageName]; |
| if (cb) { |
| return; |
| } |
| cb = new AppOpsListener(*this, packageName); |
| listener = cb; |
| } |
| |
| mAppOpsManager->startWatchingMode(AppOpsManager::OP_GET_USAGE_STATS, |
| String16(packageName.c_str()), listener); |
| } |
| |
| bool IncrementalService::unregisterAppOpsCallback(const std::string& packageName) { |
| sp<IAppOpsCallback> listener; |
| { |
| std::unique_lock lock{mCallbacksLock}; |
| auto found = mCallbackRegistered.find(packageName); |
| if (found == mCallbackRegistered.end()) { |
| return false; |
| } |
| listener = found->second; |
| mCallbackRegistered.erase(found); |
| } |
| |
| mAppOpsManager->stopWatchingMode(listener); |
| return true; |
| } |
| |
| void IncrementalService::onAppOpChanged(const std::string& packageName) { |
| if (!unregisterAppOpsCallback(packageName)) { |
| return; |
| } |
| |
| std::vector<IfsMountPtr> affected; |
| { |
| std::lock_guard l(mLock); |
| affected.reserve(mMounts.size()); |
| for (auto&& [id, ifs] : mMounts) { |
| std::unique_lock ll(ifs->lock); |
| if (ifs->mountId == id && ifs->dataLoaderStub && |
| ifs->dataLoaderStub->params().packageName == packageName) { |
| affected.push_back(ifs); |
| } |
| } |
| } |
| for (auto&& ifs : affected) { |
| std::unique_lock ll(ifs->lock); |
| disableReadLogsLocked(*ifs); |
| } |
| } |
| |
| bool IncrementalService::addTimedJob(TimedQueueWrapper& timedQueue, MountId id, Milliseconds after, |
| Job what) { |
| if (id == kInvalidStorageId) { |
| return false; |
| } |
| timedQueue.addJob(id, after, std::move(what)); |
| return true; |
| } |
| |
| bool IncrementalService::removeTimedJobs(TimedQueueWrapper& timedQueue, MountId id) { |
| if (id == kInvalidStorageId) { |
| return false; |
| } |
| timedQueue.removeJobs(id); |
| return true; |
| } |
| |
| void IncrementalService::addIfsStateCallback(StorageId storageId, IfsStateCallback callback) { |
| bool wasEmpty; |
| { |
| std::lock_guard l(mIfsStateCallbacksLock); |
| wasEmpty = mIfsStateCallbacks.empty(); |
| mIfsStateCallbacks[storageId].emplace_back(std::move(callback)); |
| } |
| if (wasEmpty) { |
| addTimedJob(*mTimedQueue, kAllStoragesId, Constants::progressUpdateInterval, |
| [this]() { processIfsStateCallbacks(); }); |
| } |
| } |
| |
| void IncrementalService::processIfsStateCallbacks() { |
| StorageId storageId = kInvalidStorageId; |
| std::vector<IfsStateCallback> local; |
| while (true) { |
| { |
| std::lock_guard l(mIfsStateCallbacksLock); |
| if (mIfsStateCallbacks.empty()) { |
| return; |
| } |
| IfsStateCallbacks::iterator it; |
| if (storageId == kInvalidStorageId) { |
| // First entry, initialize the |it|. |
| it = mIfsStateCallbacks.begin(); |
| } else { |
| // Subsequent entries, update the |storageId|, and shift to the new one (not that |
| // it guarantees much about updated items, but at least the loop will finish). |
| it = mIfsStateCallbacks.lower_bound(storageId); |
| if (it == mIfsStateCallbacks.end()) { |
| // Nothing else left, too bad. |
| break; |
| } |
| if (it->first != storageId) { |
| local.clear(); // Was removed during processing, forget the old callbacks. |
| } else { |
| // Put the 'surviving' callbacks back into the map and advance the position. |
| auto& callbacks = it->second; |
| if (callbacks.empty()) { |
| std::swap(callbacks, local); |
| } else { |
| callbacks.insert(callbacks.end(), std::move_iterator(local.begin()), |
| std::move_iterator(local.end())); |
| local.clear(); |
| } |
| if (callbacks.empty()) { |
| it = mIfsStateCallbacks.erase(it); |
| if (mIfsStateCallbacks.empty()) { |
| return; |
| } |
| } else { |
| ++it; |
| } |
| } |
| } |
| |
| if (it == mIfsStateCallbacks.end()) { |
| break; |
| } |
| |
| storageId = it->first; |
| auto& callbacks = it->second; |
| if (callbacks.empty()) { |
| // Invalid case, one extra lookup should be ok. |
| continue; |
| } |
| std::swap(callbacks, local); |
| } |
| |
| processIfsStateCallbacks(storageId, local); |
| } |
| |
| addTimedJob(*mTimedQueue, kAllStoragesId, Constants::progressUpdateInterval, |
| [this]() { processIfsStateCallbacks(); }); |
| } |
| |
| void IncrementalService::processIfsStateCallbacks(StorageId storageId, |
| std::vector<IfsStateCallback>& callbacks) { |
| const auto state = isMountFullyLoaded(storageId); |
| IfsState storageState = {}; |
| storageState.error = int(state) < 0; |
| storageState.fullyLoaded = state == incfs::LoadingState::Full; |
| if (storageState.fullyLoaded) { |
| const auto ifs = getIfs(storageId); |
| storageState.readLogsEnabled = ifs && ifs->readLogsEnabled(); |
| } |
| |
| for (auto cur = callbacks.begin(); cur != callbacks.end();) { |
| if ((*cur)(storageId, storageState)) { |
| ++cur; |
| } else { |
| cur = callbacks.erase(cur); |
| } |
| } |
| } |
| |
| void IncrementalService::removeIfsStateCallbacks(StorageId storageId) { |
| std::lock_guard l(mIfsStateCallbacksLock); |
| mIfsStateCallbacks.erase(storageId); |
| } |
| |
| void IncrementalService::getMetrics(StorageId storageId, android::os::PersistableBundle* result) { |
| const auto ifs = getIfs(storageId); |
| if (!ifs) { |
| LOG(ERROR) << "getMetrics failed, invalid storageId: " << storageId; |
| return; |
| } |
| const auto& kMetricsReadLogsEnabled = |
| os::incremental::BnIncrementalService::METRICS_READ_LOGS_ENABLED(); |
| result->putBoolean(String16(kMetricsReadLogsEnabled.c_str()), ifs->readLogsEnabled() != 0); |
| const auto incfsMetrics = mIncFs->getMetrics(ifs->metricsKey); |
| if (incfsMetrics) { |
| const auto& kMetricsTotalDelayedReads = |
| os::incremental::BnIncrementalService::METRICS_TOTAL_DELAYED_READS(); |
| const auto totalDelayedReads = |
| incfsMetrics->readsDelayedMin + incfsMetrics->readsDelayedPending; |
| result->putInt(String16(kMetricsTotalDelayedReads.c_str()), totalDelayedReads); |
| const auto& kMetricsTotalFailedReads = |
| os::incremental::BnIncrementalService::METRICS_TOTAL_FAILED_READS(); |
| const auto totalFailedReads = incfsMetrics->readsFailedTimedOut + |
| incfsMetrics->readsFailedHashVerification + incfsMetrics->readsFailedOther; |
| result->putInt(String16(kMetricsTotalFailedReads.c_str()), totalFailedReads); |
| const auto& kMetricsTotalDelayedReadsMillis = |
| os::incremental::BnIncrementalService::METRICS_TOTAL_DELAYED_READS_MILLIS(); |
| const int64_t totalDelayedReadsMillis = |
| (incfsMetrics->readsDelayedMinUs + incfsMetrics->readsDelayedPendingUs) / 1000; |
| result->putLong(String16(kMetricsTotalDelayedReadsMillis.c_str()), totalDelayedReadsMillis); |
| } |
| const auto lastReadError = mIncFs->getLastReadError(ifs->control); |
| if (lastReadError && lastReadError->timestampUs != 0) { |
| const auto& kMetricsMillisSinceLastReadError = |
| os::incremental::BnIncrementalService::METRICS_MILLIS_SINCE_LAST_READ_ERROR(); |
| result->putLong(String16(kMetricsMillisSinceLastReadError.c_str()), |
| (int64_t)elapsedUsSinceMonoTs(lastReadError->timestampUs) / 1000); |
| const auto& kMetricsLastReadErrorNo = |
| os::incremental::BnIncrementalService::METRICS_LAST_READ_ERROR_NUMBER(); |
| result->putInt(String16(kMetricsLastReadErrorNo.c_str()), lastReadError->errorNo); |
| const auto& kMetricsLastReadUid = |
| os::incremental::BnIncrementalService::METRICS_LAST_READ_ERROR_UID(); |
| result->putInt(String16(kMetricsLastReadUid.c_str()), lastReadError->uid); |
| } |
| std::unique_lock l(ifs->lock); |
| if (!ifs->dataLoaderStub) { |
| return; |
| } |
| ifs->dataLoaderStub->getMetrics(result); |
| } |
| |
| IncrementalService::DataLoaderStub::DataLoaderStub( |
| IncrementalService& service, MountId id, DataLoaderParamsParcel&& params, |
| FileSystemControlParcel&& control, DataLoaderStatusListener&& statusListener, |
| const StorageHealthCheckParams& healthCheckParams, StorageHealthListener&& healthListener, |
| std::string&& healthPath) |
| : mService(service), |
| mId(id), |
| mParams(std::move(params)), |
| mControl(std::move(control)), |
| mStatusListener(std::move(statusListener)), |
| mHealthListener(std::move(healthListener)), |
| mHealthPath(std::move(healthPath)), |
| mHealthCheckParams(healthCheckParams) { |
| if (mHealthListener && !isHealthParamsValid()) { |
| mHealthListener = {}; |
| } |
| if (!mHealthListener) { |
| // Disable advanced health check statuses. |
| mHealthCheckParams.blockedTimeoutMs = -1; |
| } |
| updateHealthStatus(); |
| } |
| |
| IncrementalService::DataLoaderStub::~DataLoaderStub() { |
| if (isValid()) { |
| cleanupResources(); |
| } |
| } |
| |
| void IncrementalService::DataLoaderStub::cleanupResources() { |
| auto now = Clock::now(); |
| { |
| std::unique_lock lock(mMutex); |
| mHealthPath.clear(); |
| unregisterFromPendingReads(); |
| resetHealthControl(); |
| mService.removeTimedJobs(*mService.mTimedQueue, mId); |
| } |
| mService.removeIfsStateCallbacks(mId); |
| |
| requestDestroy(); |
| |
| { |
| std::unique_lock lock(mMutex); |
| mParams = {}; |
| mControl = {}; |
| mHealthControl = {}; |
| mHealthListener = {}; |
| mStatusCondition.wait_until(lock, now + Constants::destroyTimeout, [this] { |
| return mCurrentStatus == IDataLoaderStatusListener::DATA_LOADER_DESTROYED; |
| }); |
| mStatusListener = {}; |
| mId = kInvalidStorageId; |
| } |
| } |
| |
| sp<content::pm::IDataLoader> IncrementalService::DataLoaderStub::getDataLoader() { |
| sp<IDataLoader> dataloader; |
| auto status = mService.mDataLoaderManager->getDataLoader(id(), &dataloader); |
| if (!status.isOk()) { |
| LOG(ERROR) << "Failed to get dataloader: " << status.toString8(); |
| return {}; |
| } |
| if (!dataloader) { |
| LOG(ERROR) << "DataLoader is null: " << status.toString8(); |
| return {}; |
| } |
| return dataloader; |
| } |
| |
| bool IncrementalService::DataLoaderStub::isSystemDataLoader() const { |
| return (params().packageName == Constants::systemPackage); |
| } |
| |
| bool IncrementalService::DataLoaderStub::requestCreate() { |
| return setTargetStatus(IDataLoaderStatusListener::DATA_LOADER_CREATED); |
| } |
| |
| bool IncrementalService::DataLoaderStub::requestStart() { |
| return setTargetStatus(IDataLoaderStatusListener::DATA_LOADER_STARTED); |
| } |
| |
| bool IncrementalService::DataLoaderStub::requestDestroy() { |
| return setTargetStatus(IDataLoaderStatusListener::DATA_LOADER_DESTROYED); |
| } |
| |
| bool IncrementalService::DataLoaderStub::setTargetStatus(int newStatus) { |
| { |
| std::unique_lock lock(mMutex); |
| setTargetStatusLocked(newStatus); |
| } |
| return fsmStep(); |
| } |
| |
| void IncrementalService::DataLoaderStub::setTargetStatusLocked(int status) { |
| auto oldStatus = mTargetStatus; |
| mTargetStatus = status; |
| mTargetStatusTs = Clock::now(); |
| LOG(DEBUG) << "Target status update for DataLoader " << id() << ": " << oldStatus << " -> " |
| << status << " (current " << mCurrentStatus << ")"; |
| } |
| |
| std::optional<Milliseconds> IncrementalService::DataLoaderStub::needToBind() { |
| std::unique_lock lock(mMutex); |
| |
| const auto now = mService.mClock->now(); |
| const bool healthy = (mPreviousBindDelay == 0ms); |
| |
| if (mCurrentStatus == IDataLoaderStatusListener::DATA_LOADER_BINDING && |
| now - mCurrentStatusTs <= Constants::bindingTimeout) { |
| LOG(INFO) << "Binding still in progress. " |
| << (healthy ? "The DL is healthy/freshly bound, ok to retry for a few times." |
| : "Already unhealthy, don't do anything.") |
| << " for storage " << mId; |
| // Binding still in progress. |
| if (!healthy) { |
| // Already unhealthy, don't do anything. |
| return {}; |
| } |
| // The DL is healthy/freshly bound, ok to retry for a few times. |
| if (now - mPreviousBindTs <= Constants::bindGracePeriod) { |
| // Still within grace period. |
| if (now - mCurrentStatusTs >= Constants::bindRetryInterval) { |
| // Retry interval passed, retrying. |
| mCurrentStatusTs = now; |
| mPreviousBindDelay = 0ms; |
| return 0ms; |
| } |
| return {}; |
| } |
| // fallthrough, mark as unhealthy, and retry with delay |
| } |
| |
| const auto previousBindTs = mPreviousBindTs; |
| mPreviousBindTs = now; |
| |
| const auto nonCrashingInterval = |
| std::max(castToMs(now - previousBindTs - mPreviousBindDelay), 100ms); |
| if (previousBindTs.time_since_epoch() == Clock::duration::zero() || |
| nonCrashingInterval > Constants::healthyDataLoaderUptime) { |
| mPreviousBindDelay = 0ms; |
| return 0ms; |
| } |
| |
| constexpr auto minBindDelayMs = castToMs(Constants::minBindDelay); |
| constexpr auto maxBindDelayMs = castToMs(Constants::maxBindDelay); |
| |
| const auto bindDelayMs = |
| std::min(std::max(mPreviousBindDelay * Constants::bindDelayMultiplier, minBindDelayMs), |
| maxBindDelayMs) |
| .count(); |
| const auto bindDelayJitterRangeMs = bindDelayMs / Constants::bindDelayJitterDivider; |
| // rand() is enough, not worth maintaining a full-blown <rand> object for delay jitter |
| const auto bindDelayJitterMs = rand() % (bindDelayJitterRangeMs * 2) - // NOLINT |
| bindDelayJitterRangeMs; |
| mPreviousBindDelay = std::chrono::milliseconds(bindDelayMs + bindDelayJitterMs); |
| return mPreviousBindDelay; |
| } |
| |
| bool IncrementalService::DataLoaderStub::bind() { |
| const auto maybeBindDelay = needToBind(); |
| if (!maybeBindDelay) { |
| LOG(DEBUG) << "Skipping bind to " << mParams.packageName << " because of pending bind."; |
| return true; |
| } |
| const auto bindDelay = *maybeBindDelay; |
| if (bindDelay > 1s) { |
| LOG(INFO) << "Delaying bind to " << mParams.packageName << " by " |
| << bindDelay.count() / 1000 << "s" |
| << " for storage " << mId; |
| } |
| |
| bool result = false; |
| auto status = mService.mDataLoaderManager->bindToDataLoader(id(), mParams, bindDelay.count(), |
| this, &result); |
| if (!status.isOk() || !result) { |
| const bool healthy = (bindDelay == 0ms); |
| LOG(ERROR) << "Failed to bind a data loader for mount " << id() |
| << (healthy ? ", retrying." : ""); |
| |
| // Internal error, retry for healthy/new DLs. |
| // Let needToBind migrate it to unhealthy after too many retries. |
| if (healthy) { |
| if (mService.addTimedJob(*mService.mTimedQueue, id(), Constants::bindRetryInterval, |
| [this]() { fsmStep(); })) { |
| // Mark as binding so that we know it's not the DL's fault. |
| setCurrentStatus(IDataLoaderStatusListener::DATA_LOADER_BINDING); |
| return true; |
| } |
| } |
| |
| return false; |
| } |
| return true; |
| } |
| |
| bool IncrementalService::DataLoaderStub::create() { |
| auto dataloader = getDataLoader(); |
| if (!dataloader) { |
| return false; |
| } |
| auto status = dataloader->create(id(), mParams, mControl, this); |
| if (!status.isOk()) { |
| LOG(ERROR) << "Failed to create DataLoader: " << status.toString8(); |
| return false; |
| } |
| return true; |
| } |
| |
| bool IncrementalService::DataLoaderStub::start() { |
| auto dataloader = getDataLoader(); |
| if (!dataloader) { |
| return false; |
| } |
| auto status = dataloader->start(id()); |
| if (!status.isOk()) { |
| LOG(ERROR) << "Failed to start DataLoader: " << status.toString8(); |
| return false; |
| } |
| return true; |
| } |
| |
| bool IncrementalService::DataLoaderStub::destroy() { |
| return mService.mDataLoaderManager->unbindFromDataLoader(id()).isOk(); |
| } |
| |
| bool IncrementalService::DataLoaderStub::fsmStep() { |
| if (!isValid()) { |
| return false; |
| } |
| |
| int currentStatus; |
| int targetStatus; |
| { |
| std::unique_lock lock(mMutex); |
| currentStatus = mCurrentStatus; |
| targetStatus = mTargetStatus; |
| } |
| |
| LOG(DEBUG) << "fsmStep: " << id() << ": " << currentStatus << " -> " << targetStatus; |
| |
| if (currentStatus == targetStatus) { |
| return true; |
| } |
| |
| switch (targetStatus) { |
| case IDataLoaderStatusListener::DATA_LOADER_DESTROYED: { |
| switch (currentStatus) { |
| case IDataLoaderStatusListener::DATA_LOADER_UNAVAILABLE: |
| case IDataLoaderStatusListener::DATA_LOADER_UNRECOVERABLE: |
| destroy(); |
| // DataLoader is broken, just assume it's destroyed. |
| compareAndSetCurrentStatus(currentStatus, |
| IDataLoaderStatusListener::DATA_LOADER_DESTROYED); |
| return true; |
| case IDataLoaderStatusListener::DATA_LOADER_BINDING: |
| compareAndSetCurrentStatus(currentStatus, |
| IDataLoaderStatusListener::DATA_LOADER_DESTROYED); |
| return true; |
| default: |
| return destroy(); |
| } |
| break; |
| } |
| case IDataLoaderStatusListener::DATA_LOADER_STARTED: { |
| switch (currentStatus) { |
| case IDataLoaderStatusListener::DATA_LOADER_CREATED: |
| case IDataLoaderStatusListener::DATA_LOADER_STOPPED: |
| return start(); |
| } |
| [[fallthrough]]; |
| } |
| case IDataLoaderStatusListener::DATA_LOADER_CREATED: |
| switch (currentStatus) { |
| case IDataLoaderStatusListener::DATA_LOADER_UNAVAILABLE: |
| case IDataLoaderStatusListener::DATA_LOADER_UNRECOVERABLE: |
| // Before binding need to make sure we are unbound. |
| // Otherwise we'll get stuck binding. |
| destroy(); |
| // DataLoader is broken, just assume it's destroyed. |
| compareAndSetCurrentStatus(currentStatus, |
| IDataLoaderStatusListener::DATA_LOADER_DESTROYED); |
| return true; |
| case IDataLoaderStatusListener::DATA_LOADER_DESTROYED: |
| case IDataLoaderStatusListener::DATA_LOADER_BINDING: |
| return bind(); |
| case IDataLoaderStatusListener::DATA_LOADER_BOUND: |
| return create(); |
| } |
| break; |
| default: |
| LOG(ERROR) << "Invalid target status: " << targetStatus |
| << ", current status: " << currentStatus; |
| break; |
| } |
| return false; |
| } |
| |
| binder::Status IncrementalService::DataLoaderStub::onStatusChanged(MountId mountId, int newStatus) { |
| if (!isValid()) { |
| if (newStatus == IDataLoaderStatusListener::DATA_LOADER_BOUND) { |
| // Async "bound" came to already destroyed stub. |
| // Unbind immediately to avoid invalid stub sitting around in DataLoaderManagerService. |
| mService.mDataLoaderManager->unbindFromDataLoader(mountId); |
| return binder::Status::ok(); |
| } |
| return binder::Status:: |
| fromServiceSpecificError(-EINVAL, "onStatusChange came to invalid DataLoaderStub"); |
| } |
| if (id() != mountId) { |
| LOG(ERROR) << "onStatusChanged: mount ID mismatch: expected " << id() |
| << ", but got: " << mountId; |
| return binder::Status::fromServiceSpecificError(-EPERM, "Mount ID mismatch."); |
| } |
| if (newStatus == IDataLoaderStatusListener::DATA_LOADER_UNAVAILABLE || |
| newStatus == IDataLoaderStatusListener::DATA_LOADER_UNRECOVERABLE) { |
| // User-provided status, let's postpone the handling to avoid possible deadlocks. |
| mService.addTimedJob(*mService.mTimedQueue, id(), Constants::userStatusDelay, |
| [this, newStatus]() { setCurrentStatus(newStatus); }); |
| return binder::Status::ok(); |
| } |
| |
| setCurrentStatus(newStatus); |
| return binder::Status::ok(); |
| } |
| |
| void IncrementalService::DataLoaderStub::setCurrentStatus(int newStatus) { |
| compareAndSetCurrentStatus(Constants::anyStatus, newStatus); |
| } |
| |
| void IncrementalService::DataLoaderStub::compareAndSetCurrentStatus(int expectedStatus, |
| int newStatus) { |
| int oldStatus, oldTargetStatus, newTargetStatus; |
| DataLoaderStatusListener listener; |
| { |
| std::unique_lock lock(mMutex); |
| if (mCurrentStatus == newStatus) { |
| return; |
| } |
| if (expectedStatus != Constants::anyStatus && expectedStatus != mCurrentStatus) { |
| return; |
| } |
| |
| oldStatus = mCurrentStatus; |
| oldTargetStatus = mTargetStatus; |
| listener = mStatusListener; |
| |
| // Change the status. |
| mCurrentStatus = newStatus; |
| mCurrentStatusTs = mService.mClock->now(); |
| |
| switch (mCurrentStatus) { |
| case IDataLoaderStatusListener::DATA_LOADER_UNAVAILABLE: |
| // Unavailable, retry. |
| setTargetStatusLocked(IDataLoaderStatusListener::DATA_LOADER_STARTED); |
| break; |
| case IDataLoaderStatusListener::DATA_LOADER_UNRECOVERABLE: |
| // Unrecoverable, just unbind. |
| setTargetStatusLocked(IDataLoaderStatusListener::DATA_LOADER_DESTROYED); |
| break; |
| default: |
| break; |
| } |
| |
| newTargetStatus = mTargetStatus; |
| } |
| |
| LOG(DEBUG) << "Current status update for DataLoader " << id() << ": " << oldStatus << " -> " |
| << newStatus << " (target " << oldTargetStatus << " -> " << newTargetStatus << ")"; |
| |
| if (listener) { |
| listener->onStatusChanged(id(), newStatus); |
| } |
| |
| fsmStep(); |
| |
| mStatusCondition.notify_all(); |
| } |
| |
| bool IncrementalService::DataLoaderStub::isHealthParamsValid() const { |
| return mHealthCheckParams.blockedTimeoutMs > 0 && |
| mHealthCheckParams.blockedTimeoutMs < mHealthCheckParams.unhealthyTimeoutMs; |
| } |
| |
| void IncrementalService::DataLoaderStub::onHealthStatus(const StorageHealthListener& healthListener, |
| int healthStatus) { |
| LOG(DEBUG) << id() << ": healthStatus: " << healthStatus; |
| if (healthListener) { |
| healthListener->onHealthStatus(id(), healthStatus); |
| } |
| mHealthStatus = healthStatus; |
| } |
| |
| void IncrementalService::DataLoaderStub::updateHealthStatus(bool baseline) { |
| LOG(DEBUG) << id() << ": updateHealthStatus" << (baseline ? " (baseline)" : ""); |
| |
| int healthStatusToReport = -1; |
| StorageHealthListener healthListener; |
| |
| { |
| std::unique_lock lock(mMutex); |
| unregisterFromPendingReads(); |
| |
| healthListener = mHealthListener; |
| |
| // Healthcheck depends on timestamp of the oldest pending read. |
| // To get it, we need to re-open a pendingReads FD to get a full list of reads. |
| // Additionally we need to re-register for epoll with fresh FDs in case there are no |
| // reads. |
| const auto now = Clock::now(); |
| const auto kernelTsUs = getOldestPendingReadTs(); |
| if (baseline) { |
| // Updating baseline only on looper/epoll callback, i.e. on new set of pending |
| // reads. |
| mHealthBase = {now, kernelTsUs}; |
| } |
| |
| if (kernelTsUs == kMaxBootClockTsUs || mHealthBase.kernelTsUs == kMaxBootClockTsUs || |
| mHealthBase.userTs > now) { |
| LOG(DEBUG) << id() << ": No pending reads or invalid base, report Ok and wait."; |
| registerForPendingReads(); |
| healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_OK; |
| lock.unlock(); |
| onHealthStatus(healthListener, healthStatusToReport); |
| return; |
| } |
| |
| resetHealthControl(); |
| |
| // Always make sure the data loader is started. |
| setTargetStatusLocked(IDataLoaderStatusListener::DATA_LOADER_STARTED); |
| |
| // Skip any further processing if health check params are invalid. |
| if (!isHealthParamsValid()) { |
| LOG(DEBUG) << id() |
| << ": Skip any further processing if health check params are invalid."; |
| healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_READS_PENDING; |
| lock.unlock(); |
| onHealthStatus(healthListener, healthStatusToReport); |
| // Triggering data loader start. This is a one-time action. |
| fsmStep(); |
| return; |
| } |
| |
| // Don't schedule timer job less than 500ms in advance. |
| static constexpr auto kTolerance = 500ms; |
| |
| const auto blockedTimeout = std::chrono::milliseconds(mHealthCheckParams.blockedTimeoutMs); |
| const auto unhealthyTimeout = |
| std::chrono::milliseconds(mHealthCheckParams.unhealthyTimeoutMs); |
| const auto unhealthyMonitoring = |
| std::max(1000ms, |
| std::chrono::milliseconds(mHealthCheckParams.unhealthyMonitoringMs)); |
| |
| const auto delta = elapsedMsSinceKernelTs(now, kernelTsUs); |
| |
| Milliseconds checkBackAfter; |
| if (delta + kTolerance < blockedTimeout) { |
| LOG(DEBUG) << id() << ": Report reads pending and wait for blocked status."; |
| checkBackAfter = blockedTimeout - delta; |
| healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_READS_PENDING; |
| } else if (delta + kTolerance < unhealthyTimeout) { |
| LOG(DEBUG) << id() << ": Report blocked and wait for unhealthy."; |
| checkBackAfter = unhealthyTimeout - delta; |
| healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_BLOCKED; |
| } else { |
| LOG(DEBUG) << id() << ": Report unhealthy and continue monitoring."; |
| checkBackAfter = unhealthyMonitoring; |
| healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_UNHEALTHY; |
| } |
| LOG(DEBUG) << id() << ": updateHealthStatus in " << double(checkBackAfter.count()) / 1000.0 |
| << "secs"; |
| mService.addTimedJob(*mService.mTimedQueue, id(), checkBackAfter, |
| [this]() { updateHealthStatus(); }); |
| } |
| |
| // With kTolerance we are expecting these to execute before the next update. |
| if (healthStatusToReport != -1) { |
| onHealthStatus(healthListener, healthStatusToReport); |
| } |
| |
| fsmStep(); |
| } |
| |
| Milliseconds IncrementalService::DataLoaderStub::elapsedMsSinceKernelTs(TimePoint now, |
| BootClockTsUs kernelTsUs) { |
| const auto kernelDeltaUs = kernelTsUs - mHealthBase.kernelTsUs; |
| const auto userTs = mHealthBase.userTs + std::chrono::microseconds(kernelDeltaUs); |
| return std::chrono::duration_cast<Milliseconds>(now - userTs); |
| } |
| |
| const incfs::UniqueControl& IncrementalService::DataLoaderStub::initializeHealthControl() { |
| if (mHealthPath.empty()) { |
| resetHealthControl(); |
| return mHealthControl; |
| } |
| if (mHealthControl.pendingReads() < 0) { |
| mHealthControl = mService.mIncFs->openMount(mHealthPath); |
| } |
| if (mHealthControl.pendingReads() < 0) { |
| LOG(ERROR) << "Failed to open health control for: " << id() << ", path: " << mHealthPath |
| << "(" << mHealthControl.cmd() << ":" << mHealthControl.pendingReads() << ":" |
| << mHealthControl.logs() << ")"; |
| } |
| return mHealthControl; |
| } |
| |
| void IncrementalService::DataLoaderStub::resetHealthControl() { |
| mHealthControl = {}; |
| } |
| |
| BootClockTsUs IncrementalService::DataLoaderStub::getOldestPendingReadTs() { |
| auto result = kMaxBootClockTsUs; |
| |
| const auto& control = initializeHealthControl(); |
| if (control.pendingReads() < 0) { |
| return result; |
| } |
| |
| if (mService.mIncFs->waitForPendingReads(control, 0ms, &mLastPendingReads) != |
| android::incfs::WaitResult::HaveData || |
| mLastPendingReads.empty()) { |
| // Clear previous pending reads |
| mLastPendingReads.clear(); |
| return result; |
| } |
| |
| LOG(DEBUG) << id() << ": pendingReads: fd(" << control.pendingReads() << "), count(" |
| << mLastPendingReads.size() << "), block: " << mLastPendingReads.front().block |
| << ", time: " << mLastPendingReads.front().bootClockTsUs |
| << ", uid: " << mLastPendingReads.front().uid; |
| |
| return getOldestTsFromLastPendingReads(); |
| } |
| |
| void IncrementalService::DataLoaderStub::registerForPendingReads() { |
| const auto pendingReadsFd = mHealthControl.pendingReads(); |
| if (pendingReadsFd < 0) { |
| return; |
| } |
| |
| LOG(DEBUG) << id() << ": addFd(pendingReadsFd): " << pendingReadsFd; |
| |
| mService.mLooper->addFd( |
| pendingReadsFd, android::Looper::POLL_CALLBACK, android::Looper::EVENT_INPUT, |
| [](int, int, void* data) -> int { |
| auto self = (DataLoaderStub*)data; |
| self->updateHealthStatus(/*baseline=*/true); |
| return 0; |
| }, |
| this); |
| mService.mLooper->wake(); |
| } |
| |
| BootClockTsUs IncrementalService::DataLoaderStub::getOldestTsFromLastPendingReads() { |
| auto result = kMaxBootClockTsUs; |
| for (auto&& pendingRead : mLastPendingReads) { |
| result = std::min(result, pendingRead.bootClockTsUs); |
| } |
| return result; |
| } |
| |
| void IncrementalService::DataLoaderStub::getMetrics(android::os::PersistableBundle* result) { |
| const auto duration = elapsedMsSinceOldestPendingRead(); |
| if (duration >= 0) { |
| const auto& kMetricsMillisSinceOldestPendingRead = |
| os::incremental::BnIncrementalService::METRICS_MILLIS_SINCE_OLDEST_PENDING_READ(); |
| result->putLong(String16(kMetricsMillisSinceOldestPendingRead.c_str()), duration); |
| } |
| const auto& kMetricsStorageHealthStatusCode = |
| os::incremental::BnIncrementalService::METRICS_STORAGE_HEALTH_STATUS_CODE(); |
| result->putInt(String16(kMetricsStorageHealthStatusCode.c_str()), mHealthStatus); |
| const auto& kMetricsDataLoaderStatusCode = |
| os::incremental::BnIncrementalService::METRICS_DATA_LOADER_STATUS_CODE(); |
| result->putInt(String16(kMetricsDataLoaderStatusCode.c_str()), mCurrentStatus); |
| const auto& kMetricsMillisSinceLastDataLoaderBind = |
| os::incremental::BnIncrementalService::METRICS_MILLIS_SINCE_LAST_DATA_LOADER_BIND(); |
| result->putLong(String16(kMetricsMillisSinceLastDataLoaderBind.c_str()), |
| elapsedMcs(mPreviousBindTs, mService.mClock->now()) / 1000); |
| const auto& kMetricsDataLoaderBindDelayMillis = |
| os::incremental::BnIncrementalService::METRICS_DATA_LOADER_BIND_DELAY_MILLIS(); |
| result->putLong(String16(kMetricsDataLoaderBindDelayMillis.c_str()), |
| mPreviousBindDelay.count()); |
| } |
| |
| long IncrementalService::DataLoaderStub::elapsedMsSinceOldestPendingRead() { |
| const auto oldestPendingReadKernelTs = getOldestTsFromLastPendingReads(); |
| if (oldestPendingReadKernelTs == kMaxBootClockTsUs) { |
| return 0; |
| } |
| return elapsedMsSinceKernelTs(Clock::now(), oldestPendingReadKernelTs).count(); |
| } |
| |
| void IncrementalService::DataLoaderStub::unregisterFromPendingReads() { |
| const auto pendingReadsFd = mHealthControl.pendingReads(); |
| if (pendingReadsFd < 0) { |
| return; |
| } |
| |
| LOG(DEBUG) << id() << ": removeFd(pendingReadsFd): " << pendingReadsFd; |
| |
| mService.mLooper->removeFd(pendingReadsFd); |
| mService.mLooper->wake(); |
| } |
| |
| void IncrementalService::DataLoaderStub::setHealthListener( |
| const StorageHealthCheckParams& healthCheckParams, StorageHealthListener&& healthListener) { |
| std::lock_guard lock(mMutex); |
| mHealthCheckParams = healthCheckParams; |
| mHealthListener = std::move(healthListener); |
| if (!mHealthListener) { |
| mHealthCheckParams.blockedTimeoutMs = -1; |
| } |
| } |
| |
| static std::string toHexString(const RawMetadata& metadata) { |
| int n = metadata.size(); |
| std::string res(n * 2, '\0'); |
| // Same as incfs::toString(fileId) |
| static constexpr char kHexChar[] = "0123456789abcdef"; |
| for (int i = 0; i < n; ++i) { |
| res[i * 2] = kHexChar[(metadata[i] & 0xf0) >> 4]; |
| res[i * 2 + 1] = kHexChar[(metadata[i] & 0x0f)]; |
| } |
| return res; |
| } |
| |
| void IncrementalService::DataLoaderStub::onDump(int fd) { |
| dprintf(fd, " dataLoader: {\n"); |
| dprintf(fd, " currentStatus: %d\n", mCurrentStatus); |
| dprintf(fd, " currentStatusTs: %lldmcs\n", |
| (long long)(elapsedMcs(mCurrentStatusTs, Clock::now()))); |
| dprintf(fd, " targetStatus: %d\n", mTargetStatus); |
| dprintf(fd, " targetStatusTs: %lldmcs\n", |
| (long long)(elapsedMcs(mTargetStatusTs, Clock::now()))); |
| dprintf(fd, " health: {\n"); |
| dprintf(fd, " path: %s\n", mHealthPath.c_str()); |
| dprintf(fd, " base: %lldmcs (%lld)\n", |
| (long long)(elapsedMcs(mHealthBase.userTs, Clock::now())), |
| (long long)mHealthBase.kernelTsUs); |
| dprintf(fd, " blockedTimeoutMs: %d\n", int(mHealthCheckParams.blockedTimeoutMs)); |
| dprintf(fd, " unhealthyTimeoutMs: %d\n", int(mHealthCheckParams.unhealthyTimeoutMs)); |
| dprintf(fd, " unhealthyMonitoringMs: %d\n", |
| int(mHealthCheckParams.unhealthyMonitoringMs)); |
| dprintf(fd, " lastPendingReads: \n"); |
| const auto control = mService.mIncFs->openMount(mHealthPath); |
| for (auto&& pendingRead : mLastPendingReads) { |
| dprintf(fd, " fileId: %s\n", IncFsWrapper::toString(pendingRead.id).c_str()); |
| const auto metadata = mService.mIncFs->getMetadata(control, pendingRead.id); |
| dprintf(fd, " metadataHex: %s\n", toHexString(metadata).c_str()); |
| dprintf(fd, " blockIndex: %d\n", pendingRead.block); |
| dprintf(fd, " bootClockTsUs: %lld\n", (long long)pendingRead.bootClockTsUs); |
| } |
| dprintf(fd, " bind: %llds ago (delay: %llds)\n", |
| (long long)(elapsedMcs(mPreviousBindTs, mService.mClock->now()) / 1000000), |
| (long long)(mPreviousBindDelay.count() / 1000)); |
| dprintf(fd, " }\n"); |
| const auto& params = mParams; |
| dprintf(fd, " dataLoaderParams: {\n"); |
| dprintf(fd, " type: %s\n", toString(params.type).c_str()); |
| dprintf(fd, " packageName: %s\n", params.packageName.c_str()); |
| dprintf(fd, " className: %s\n", params.className.c_str()); |
| dprintf(fd, " arguments: %s\n", params.arguments.c_str()); |
| dprintf(fd, " }\n"); |
| dprintf(fd, " }\n"); |
| } |
| |
| void IncrementalService::AppOpsListener::opChanged(int32_t, const String16&) { |
| incrementalService.onAppOpChanged(packageName); |
| } |
| |
| binder::Status IncrementalService::IncrementalServiceConnector::setStorageParams( |
| bool enableReadLogs, int32_t* _aidl_return) { |
| *_aidl_return = incrementalService.setStorageParams(storage, enableReadLogs); |
| return binder::Status::ok(); |
| } |
| |
| FileId IncrementalService::idFromMetadata(std::span<const uint8_t> metadata) { |
| return IncFs_FileIdFromMetadata({(const char*)metadata.data(), (IncFsSize)metadata.size()}); |
| } |
| |
| } // namespace android::incremental |