6e0b0fa671
10secs should be more than enough to stop whatever the DL is doing. Besides, DL has its own set of IncFS FDs and should not crash. Bug: 189222575 Fixes: 189222575 Test: atest IncrementalServiceTest Change-Id: I73cb27d61c7418adeea7536c8263e6ba8c77fd3e
3190 lines
122 KiB
C++
3190 lines
122 KiB
C++
/*
|
|
* Copyright (C) 2019 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#define LOG_TAG "IncrementalService"
|
|
|
|
#include "IncrementalService.h"
|
|
|
|
#include <android-base/logging.h>
|
|
#include <android-base/no_destructor.h>
|
|
#include <android-base/properties.h>
|
|
#include <android-base/stringprintf.h>
|
|
#include <binder/AppOpsManager.h>
|
|
#include <binder/Status.h>
|
|
#include <sys/stat.h>
|
|
#include <uuid/uuid.h>
|
|
|
|
#include <charconv>
|
|
#include <ctime>
|
|
#include <iterator>
|
|
#include <span>
|
|
#include <type_traits>
|
|
|
|
#include "IncrementalServiceValidation.h"
|
|
#include "Metadata.pb.h"
|
|
|
|
using namespace std::literals;
|
|
|
|
constexpr const char* kLoaderUsageStats = "android.permission.LOADER_USAGE_STATS";
|
|
constexpr const char* kOpUsage = "android:loader_usage_stats";
|
|
|
|
constexpr const char* kInteractAcrossUsers = "android.permission.INTERACT_ACROSS_USERS";
|
|
|
|
namespace android::incremental {
|
|
|
|
using content::pm::DataLoaderParamsParcel;
|
|
using content::pm::FileSystemControlParcel;
|
|
using content::pm::IDataLoader;
|
|
|
|
namespace {
|
|
|
|
using IncrementalFileSystemControlParcel = os::incremental::IncrementalFileSystemControlParcel;
|
|
|
|
struct Constants {
|
|
static constexpr auto backing = "backing_store"sv;
|
|
static constexpr auto mount = "mount"sv;
|
|
static constexpr auto mountKeyPrefix = "MT_"sv;
|
|
static constexpr auto storagePrefix = "st"sv;
|
|
static constexpr auto mountpointMdPrefix = ".mountpoint."sv;
|
|
static constexpr auto infoMdName = ".info"sv;
|
|
static constexpr auto readLogsDisabledMarkerName = ".readlogs_disabled"sv;
|
|
static constexpr auto libDir = "lib"sv;
|
|
static constexpr auto libSuffix = ".so"sv;
|
|
static constexpr auto blockSize = 4096;
|
|
static constexpr auto systemPackage = "android"sv;
|
|
|
|
static constexpr auto userStatusDelay = 100ms;
|
|
|
|
static constexpr auto progressUpdateInterval = 1000ms;
|
|
static constexpr auto perUidTimeoutOffset = progressUpdateInterval * 2;
|
|
static constexpr auto minPerUidTimeout = progressUpdateInterval * 3;
|
|
|
|
// If DL was up and not crashing for 10mins, we consider it healthy and reset all delays.
|
|
static constexpr auto healthyDataLoaderUptime = 10min;
|
|
|
|
// For healthy DLs, we'll retry every ~5secs for ~10min
|
|
static constexpr auto bindRetryInterval = 5s;
|
|
static constexpr auto bindGracePeriod = 10min;
|
|
|
|
static constexpr auto bindingTimeout = 1min;
|
|
|
|
// 1s, 10s, 100s (~2min), 1000s (~15min), 10000s (~3hrs)
|
|
static constexpr auto minBindDelay = 1s;
|
|
static constexpr auto maxBindDelay = 10000s;
|
|
static constexpr auto bindDelayMultiplier = 10;
|
|
static constexpr auto bindDelayJitterDivider = 10;
|
|
|
|
// Max interval after system invoked the DL when readlog collection can be enabled.
|
|
static constexpr auto readLogsMaxInterval = 2h;
|
|
|
|
// How long should we wait till dataLoader reports destroyed.
|
|
static constexpr auto destroyTimeout = 10s;
|
|
|
|
static constexpr auto anyStatus = INT_MIN;
|
|
};
|
|
|
|
static const Constants& constants() {
|
|
static constexpr Constants c;
|
|
return c;
|
|
}
|
|
|
|
static bool isPageAligned(IncFsSize s) {
|
|
return (s & (Constants::blockSize - 1)) == 0;
|
|
}
|
|
|
|
static bool getAlwaysEnableReadTimeoutsForSystemDataLoaders() {
|
|
return android::base::
|
|
GetBoolProperty("debug.incremental.always_enable_read_timeouts_for_system_dataloaders",
|
|
true);
|
|
}
|
|
|
|
static bool getEnforceReadLogsMaxIntervalForSystemDataLoaders() {
|
|
return android::base::GetBoolProperty("debug.incremental.enforce_readlogs_max_interval_for_"
|
|
"system_dataloaders",
|
|
false);
|
|
}
|
|
|
|
static Seconds getReadLogsMaxInterval() {
|
|
constexpr int limit = duration_cast<Seconds>(Constants::readLogsMaxInterval).count();
|
|
int readlogs_max_interval_secs =
|
|
std::min(limit,
|
|
android::base::GetIntProperty<
|
|
int>("debug.incremental.readlogs_max_interval_sec", limit));
|
|
return Seconds{readlogs_max_interval_secs};
|
|
}
|
|
|
|
template <base::LogSeverity level = base::ERROR>
|
|
bool mkdirOrLog(std::string_view name, int mode = 0770, bool allowExisting = true) {
|
|
auto cstr = path::c_str(name);
|
|
if (::mkdir(cstr, mode)) {
|
|
if (!allowExisting || errno != EEXIST) {
|
|
PLOG(level) << "Can't create directory '" << name << '\'';
|
|
return false;
|
|
}
|
|
struct stat st;
|
|
if (::stat(cstr, &st) || !S_ISDIR(st.st_mode)) {
|
|
PLOG(level) << "Path exists but is not a directory: '" << name << '\'';
|
|
return false;
|
|
}
|
|
}
|
|
if (::chmod(cstr, mode)) {
|
|
PLOG(level) << "Changing permission failed for '" << name << '\'';
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static std::string toMountKey(std::string_view path) {
|
|
if (path.empty()) {
|
|
return "@none";
|
|
}
|
|
if (path == "/"sv) {
|
|
return "@root";
|
|
}
|
|
if (path::isAbsolute(path)) {
|
|
path.remove_prefix(1);
|
|
}
|
|
if (path.size() > 16) {
|
|
path = path.substr(0, 16);
|
|
}
|
|
std::string res(path);
|
|
std::replace_if(
|
|
res.begin(), res.end(), [](char c) { return c == '/' || c == '@'; }, '_');
|
|
return std::string(constants().mountKeyPrefix) += res;
|
|
}
|
|
|
|
static std::pair<std::string, std::string> makeMountDir(std::string_view incrementalDir,
|
|
std::string_view path) {
|
|
auto mountKey = toMountKey(path);
|
|
const auto prefixSize = mountKey.size();
|
|
for (int counter = 0; counter < 1000;
|
|
mountKey.resize(prefixSize), base::StringAppendF(&mountKey, "%d", counter++)) {
|
|
auto mountRoot = path::join(incrementalDir, mountKey);
|
|
if (mkdirOrLog(mountRoot, 0777, false)) {
|
|
return {mountKey, mountRoot};
|
|
}
|
|
}
|
|
return {};
|
|
}
|
|
|
|
template <class Map>
|
|
typename Map::const_iterator findParentPath(const Map& map, std::string_view path) {
|
|
const auto nextIt = map.upper_bound(path);
|
|
if (nextIt == map.begin()) {
|
|
return map.end();
|
|
}
|
|
const auto suspectIt = std::prev(nextIt);
|
|
if (!path::startsWith(path, suspectIt->first)) {
|
|
return map.end();
|
|
}
|
|
return suspectIt;
|
|
}
|
|
|
|
static base::unique_fd dup(base::borrowed_fd fd) {
|
|
const auto res = fcntl(fd.get(), F_DUPFD_CLOEXEC, 0);
|
|
return base::unique_fd(res);
|
|
}
|
|
|
|
template <class ProtoMessage, class Control>
|
|
static ProtoMessage parseFromIncfs(const IncFsWrapper* incfs, const Control& control,
|
|
std::string_view path) {
|
|
auto md = incfs->getMetadata(control, path);
|
|
ProtoMessage message;
|
|
return message.ParseFromArray(md.data(), md.size()) ? message : ProtoMessage{};
|
|
}
|
|
|
|
static bool isValidMountTarget(std::string_view path) {
|
|
return path::isAbsolute(path) && path::isEmptyDir(path).value_or(true);
|
|
}
|
|
|
|
std::string makeUniqueName(std::string_view prefix) {
|
|
static constexpr auto uuidStringSize = 36;
|
|
|
|
uuid_t guid;
|
|
uuid_generate(guid);
|
|
|
|
std::string name;
|
|
const auto prefixSize = prefix.size();
|
|
name.reserve(prefixSize + uuidStringSize);
|
|
|
|
name = prefix;
|
|
name.resize(prefixSize + uuidStringSize);
|
|
uuid_unparse(guid, name.data() + prefixSize);
|
|
|
|
return name;
|
|
}
|
|
|
|
std::string makeBindMdName() {
|
|
return makeUniqueName(constants().mountpointMdPrefix);
|
|
}
|
|
|
|
static bool checkReadLogsDisabledMarker(std::string_view root) {
|
|
const auto markerPath = path::c_str(path::join(root, constants().readLogsDisabledMarkerName));
|
|
struct stat st;
|
|
return (::stat(markerPath, &st) == 0);
|
|
}
|
|
|
|
} // namespace
|
|
|
|
IncrementalService::IncFsMount::~IncFsMount() {
|
|
if (dataLoaderStub) {
|
|
dataLoaderStub->cleanupResources();
|
|
dataLoaderStub = {};
|
|
}
|
|
control.close();
|
|
LOG(INFO) << "Unmounting and cleaning up mount " << mountId << " with root '" << root << '\'';
|
|
for (auto&& [target, _] : bindPoints) {
|
|
LOG(INFO) << " bind: " << target;
|
|
incrementalService.mVold->unmountIncFs(target);
|
|
}
|
|
LOG(INFO) << " root: " << root;
|
|
incrementalService.mVold->unmountIncFs(path::join(root, constants().mount));
|
|
cleanupFilesystem(root);
|
|
}
|
|
|
|
auto IncrementalService::IncFsMount::makeStorage(StorageId id) -> StorageMap::iterator {
|
|
std::string name;
|
|
for (int no = nextStorageDirNo.fetch_add(1, std::memory_order_relaxed), i = 0;
|
|
i < 1024 && no >= 0; no = nextStorageDirNo.fetch_add(1, std::memory_order_relaxed), ++i) {
|
|
name.clear();
|
|
base::StringAppendF(&name, "%.*s_%d_%d", int(constants().storagePrefix.size()),
|
|
constants().storagePrefix.data(), id, no);
|
|
auto fullName = path::join(root, constants().mount, name);
|
|
if (auto err = incrementalService.mIncFs->makeDir(control, fullName, 0755); !err) {
|
|
std::lock_guard l(lock);
|
|
return storages.insert_or_assign(id, Storage{std::move(fullName)}).first;
|
|
} else if (err != EEXIST) {
|
|
LOG(ERROR) << __func__ << "(): failed to create dir |" << fullName << "| " << err;
|
|
break;
|
|
}
|
|
}
|
|
nextStorageDirNo = 0;
|
|
return storages.end();
|
|
}
|
|
|
|
template <class Func>
|
|
static auto makeCleanup(Func&& f) requires(!std::is_lvalue_reference_v<Func>) {
|
|
// ok to move a 'forwarding' reference here as lvalues are disabled anyway
|
|
auto deleter = [f = std::move(f)](auto) { // NOLINT
|
|
f();
|
|
};
|
|
// &f is a dangling pointer here, but we actually never use it as deleter moves it in.
|
|
return std::unique_ptr<Func, decltype(deleter)>(&f, std::move(deleter));
|
|
}
|
|
|
|
static auto openDir(const char* dir) {
|
|
struct DirCloser {
|
|
void operator()(DIR* d) const noexcept { ::closedir(d); }
|
|
};
|
|
return std::unique_ptr<DIR, DirCloser>(::opendir(dir));
|
|
}
|
|
|
|
static auto openDir(std::string_view dir) {
|
|
return openDir(path::c_str(dir));
|
|
}
|
|
|
|
static int rmDirContent(const char* path) {
|
|
auto dir = openDir(path);
|
|
if (!dir) {
|
|
return -EINVAL;
|
|
}
|
|
while (auto entry = ::readdir(dir.get())) {
|
|
if (entry->d_name == "."sv || entry->d_name == ".."sv) {
|
|
continue;
|
|
}
|
|
auto fullPath = base::StringPrintf("%s/%s", path, entry->d_name);
|
|
if (entry->d_type == DT_DIR) {
|
|
if (const auto err = rmDirContent(fullPath.c_str()); err != 0) {
|
|
PLOG(WARNING) << "Failed to delete " << fullPath << " content";
|
|
return err;
|
|
}
|
|
if (const auto err = ::rmdir(fullPath.c_str()); err != 0) {
|
|
PLOG(WARNING) << "Failed to rmdir " << fullPath;
|
|
return err;
|
|
}
|
|
} else {
|
|
if (const auto err = ::unlink(fullPath.c_str()); err != 0) {
|
|
PLOG(WARNING) << "Failed to delete " << fullPath;
|
|
return err;
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
void IncrementalService::IncFsMount::cleanupFilesystem(std::string_view root) {
|
|
rmDirContent(path::join(root, constants().backing).c_str());
|
|
::rmdir(path::join(root, constants().backing).c_str());
|
|
::rmdir(path::join(root, constants().mount).c_str());
|
|
::rmdir(path::c_str(root));
|
|
}
|
|
|
|
void IncrementalService::IncFsMount::setFlag(StorageFlags flag, bool value) {
|
|
if (value) {
|
|
flags |= flag;
|
|
} else {
|
|
flags &= ~flag;
|
|
}
|
|
}
|
|
|
|
IncrementalService::IncrementalService(ServiceManagerWrapper&& sm, std::string_view rootDir)
|
|
: mVold(sm.getVoldService()),
|
|
mDataLoaderManager(sm.getDataLoaderManager()),
|
|
mIncFs(sm.getIncFs()),
|
|
mAppOpsManager(sm.getAppOpsManager()),
|
|
mJni(sm.getJni()),
|
|
mLooper(sm.getLooper()),
|
|
mTimedQueue(sm.getTimedQueue()),
|
|
mProgressUpdateJobQueue(sm.getProgressUpdateJobQueue()),
|
|
mFs(sm.getFs()),
|
|
mClock(sm.getClock()),
|
|
mIncrementalDir(rootDir) {
|
|
CHECK(mVold) << "Vold service is unavailable";
|
|
CHECK(mDataLoaderManager) << "DataLoaderManagerService is unavailable";
|
|
CHECK(mAppOpsManager) << "AppOpsManager is unavailable";
|
|
CHECK(mJni) << "JNI is unavailable";
|
|
CHECK(mLooper) << "Looper is unavailable";
|
|
CHECK(mTimedQueue) << "TimedQueue is unavailable";
|
|
CHECK(mProgressUpdateJobQueue) << "mProgressUpdateJobQueue is unavailable";
|
|
CHECK(mFs) << "Fs is unavailable";
|
|
CHECK(mClock) << "Clock is unavailable";
|
|
|
|
mJobQueue.reserve(16);
|
|
mJobProcessor = std::thread([this]() {
|
|
mJni->initializeForCurrentThread();
|
|
runJobProcessing();
|
|
});
|
|
mCmdLooperThread = std::thread([this]() {
|
|
mJni->initializeForCurrentThread();
|
|
runCmdLooper();
|
|
});
|
|
|
|
const auto mountedRootNames = adoptMountedInstances();
|
|
mountExistingImages(mountedRootNames);
|
|
}
|
|
|
|
IncrementalService::~IncrementalService() {
|
|
{
|
|
std::lock_guard lock(mJobMutex);
|
|
mRunning = false;
|
|
}
|
|
mJobCondition.notify_all();
|
|
mJobProcessor.join();
|
|
mLooper->wake();
|
|
mCmdLooperThread.join();
|
|
mTimedQueue->stop();
|
|
mProgressUpdateJobQueue->stop();
|
|
// Ensure that mounts are destroyed while the service is still valid.
|
|
mBindsByPath.clear();
|
|
mMounts.clear();
|
|
}
|
|
|
|
static const char* toString(IncrementalService::BindKind kind) {
|
|
switch (kind) {
|
|
case IncrementalService::BindKind::Temporary:
|
|
return "Temporary";
|
|
case IncrementalService::BindKind::Permanent:
|
|
return "Permanent";
|
|
}
|
|
}
|
|
|
|
template <class Duration>
|
|
static int64_t elapsedMcs(Duration start, Duration end) {
|
|
return std::chrono::duration_cast<std::chrono::microseconds>(end - start).count();
|
|
}
|
|
|
|
int64_t IncrementalService::elapsedUsSinceMonoTs(uint64_t monoTsUs) {
|
|
const auto now = mClock->now();
|
|
const auto nowUs = static_cast<uint64_t>(
|
|
duration_cast<std::chrono::microseconds>(now.time_since_epoch()).count());
|
|
return nowUs - monoTsUs;
|
|
}
|
|
|
|
void IncrementalService::onDump(int fd) {
|
|
dprintf(fd, "Incremental is %s\n", incfs::enabled() ? "ENABLED" : "DISABLED");
|
|
dprintf(fd, "IncFs features: 0x%x\n", int(mIncFs->features()));
|
|
dprintf(fd, "Incremental dir: %s\n", mIncrementalDir.c_str());
|
|
|
|
std::unique_lock l(mLock);
|
|
|
|
dprintf(fd, "Mounts (%d): {\n", int(mMounts.size()));
|
|
for (auto&& [id, ifs] : mMounts) {
|
|
std::unique_lock ll(ifs->lock);
|
|
const IncFsMount& mnt = *ifs;
|
|
dprintf(fd, " [%d]: {\n", id);
|
|
if (id != mnt.mountId) {
|
|
dprintf(fd, " reference to mountId: %d\n", mnt.mountId);
|
|
} else {
|
|
dprintf(fd, " mountId: %d\n", mnt.mountId);
|
|
dprintf(fd, " root: %s\n", mnt.root.c_str());
|
|
const auto& metricsInstanceName = ifs->metricsKey;
|
|
dprintf(fd, " metrics instance name: %s\n", path::c_str(metricsInstanceName).get());
|
|
dprintf(fd, " nextStorageDirNo: %d\n", mnt.nextStorageDirNo.load());
|
|
dprintf(fd, " flags: %d\n", int(mnt.flags));
|
|
if (mnt.startLoadingTs.time_since_epoch() == Clock::duration::zero()) {
|
|
dprintf(fd, " not loading\n");
|
|
} else {
|
|
dprintf(fd, " startLoading: %llds\n",
|
|
(long long)(elapsedMcs(mnt.startLoadingTs, Clock::now()) / 1000000));
|
|
}
|
|
if (mnt.dataLoaderStub) {
|
|
mnt.dataLoaderStub->onDump(fd);
|
|
} else {
|
|
dprintf(fd, " dataLoader: null\n");
|
|
}
|
|
dprintf(fd, " storages (%d): {\n", int(mnt.storages.size()));
|
|
for (auto&& [storageId, storage] : mnt.storages) {
|
|
dprintf(fd, " [%d] -> [%s] (%d %% loaded) \n", storageId, storage.name.c_str(),
|
|
(int)(getLoadingProgressFromPath(mnt, storage.name.c_str()).getProgress() *
|
|
100));
|
|
}
|
|
dprintf(fd, " }\n");
|
|
|
|
dprintf(fd, " bindPoints (%d): {\n", int(mnt.bindPoints.size()));
|
|
for (auto&& [target, bind] : mnt.bindPoints) {
|
|
dprintf(fd, " [%s]->[%d]:\n", target.c_str(), bind.storage);
|
|
dprintf(fd, " savedFilename: %s\n", bind.savedFilename.c_str());
|
|
dprintf(fd, " sourceDir: %s\n", bind.sourceDir.c_str());
|
|
dprintf(fd, " kind: %s\n", toString(bind.kind));
|
|
}
|
|
dprintf(fd, " }\n");
|
|
|
|
dprintf(fd, " incfsMetrics: {\n");
|
|
const auto incfsMetrics = mIncFs->getMetrics(metricsInstanceName);
|
|
if (incfsMetrics) {
|
|
dprintf(fd, " readsDelayedMin: %d\n", incfsMetrics.value().readsDelayedMin);
|
|
dprintf(fd, " readsDelayedMinUs: %lld\n",
|
|
(long long)incfsMetrics.value().readsDelayedMinUs);
|
|
dprintf(fd, " readsDelayedPending: %d\n",
|
|
incfsMetrics.value().readsDelayedPending);
|
|
dprintf(fd, " readsDelayedPendingUs: %lld\n",
|
|
(long long)incfsMetrics.value().readsDelayedPendingUs);
|
|
dprintf(fd, " readsFailedHashVerification: %d\n",
|
|
incfsMetrics.value().readsFailedHashVerification);
|
|
dprintf(fd, " readsFailedOther: %d\n", incfsMetrics.value().readsFailedOther);
|
|
dprintf(fd, " readsFailedTimedOut: %d\n",
|
|
incfsMetrics.value().readsFailedTimedOut);
|
|
} else {
|
|
dprintf(fd, " Metrics not available. Errno: %d\n", errno);
|
|
}
|
|
dprintf(fd, " }\n");
|
|
|
|
const auto lastReadError = mIncFs->getLastReadError(ifs->control);
|
|
const auto errorNo = errno;
|
|
dprintf(fd, " lastReadError: {\n");
|
|
if (lastReadError) {
|
|
if (lastReadError->timestampUs == 0) {
|
|
dprintf(fd, " No read errors.\n");
|
|
} else {
|
|
dprintf(fd, " fileId: %s\n",
|
|
IncFsWrapper::toString(lastReadError->id).c_str());
|
|
dprintf(fd, " time: %llu microseconds ago\n",
|
|
(unsigned long long)elapsedUsSinceMonoTs(lastReadError->timestampUs));
|
|
dprintf(fd, " blockIndex: %d\n", lastReadError->block);
|
|
dprintf(fd, " errno: %d\n", lastReadError->errorNo);
|
|
}
|
|
} else {
|
|
dprintf(fd, " Info not available. Errno: %d\n", errorNo);
|
|
}
|
|
dprintf(fd, " }\n");
|
|
}
|
|
dprintf(fd, " }\n");
|
|
}
|
|
dprintf(fd, "}\n");
|
|
dprintf(fd, "Sorted binds (%d): {\n", int(mBindsByPath.size()));
|
|
for (auto&& [target, mountPairIt] : mBindsByPath) {
|
|
const auto& bind = mountPairIt->second;
|
|
dprintf(fd, " [%s]->[%d]:\n", target.c_str(), bind.storage);
|
|
dprintf(fd, " savedFilename: %s\n", bind.savedFilename.c_str());
|
|
dprintf(fd, " sourceDir: %s\n", bind.sourceDir.c_str());
|
|
dprintf(fd, " kind: %s\n", toString(bind.kind));
|
|
}
|
|
dprintf(fd, "}\n");
|
|
}
|
|
|
|
bool IncrementalService::needStartDataLoaderLocked(IncFsMount& ifs) {
|
|
if (!ifs.dataLoaderStub) {
|
|
return false;
|
|
}
|
|
if (ifs.dataLoaderStub->isSystemDataLoader()) {
|
|
return true;
|
|
}
|
|
|
|
return mIncFs->isEverythingFullyLoaded(ifs.control) == incfs::LoadingState::MissingBlocks;
|
|
}
|
|
|
|
void IncrementalService::onSystemReady() {
|
|
if (mSystemReady.exchange(true)) {
|
|
return;
|
|
}
|
|
|
|
std::vector<IfsMountPtr> mounts;
|
|
{
|
|
std::lock_guard l(mLock);
|
|
mounts.reserve(mMounts.size());
|
|
for (auto&& [id, ifs] : mMounts) {
|
|
std::unique_lock ll(ifs->lock);
|
|
|
|
if (ifs->mountId != id) {
|
|
continue;
|
|
}
|
|
|
|
if (needStartDataLoaderLocked(*ifs)) {
|
|
mounts.push_back(ifs);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (mounts.empty()) {
|
|
return;
|
|
}
|
|
|
|
std::thread([this, mounts = std::move(mounts)]() {
|
|
mJni->initializeForCurrentThread();
|
|
for (auto&& ifs : mounts) {
|
|
std::unique_lock l(ifs->lock);
|
|
if (ifs->dataLoaderStub) {
|
|
ifs->dataLoaderStub->requestStart();
|
|
}
|
|
}
|
|
}).detach();
|
|
}
|
|
|
|
auto IncrementalService::getStorageSlotLocked() -> MountMap::iterator {
|
|
for (;;) {
|
|
if (mNextId == kMaxStorageId) {
|
|
mNextId = 0;
|
|
}
|
|
auto id = ++mNextId;
|
|
auto [it, inserted] = mMounts.try_emplace(id, nullptr);
|
|
if (inserted) {
|
|
return it;
|
|
}
|
|
}
|
|
}
|
|
|
|
StorageId IncrementalService::createStorage(std::string_view mountPoint,
|
|
content::pm::DataLoaderParamsParcel dataLoaderParams,
|
|
CreateOptions options) {
|
|
LOG(INFO) << "createStorage: " << mountPoint << " | " << int(options);
|
|
if (!path::isAbsolute(mountPoint)) {
|
|
LOG(ERROR) << "path is not absolute: " << mountPoint;
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
auto mountNorm = path::normalize(mountPoint);
|
|
{
|
|
const auto id = findStorageId(mountNorm);
|
|
if (id != kInvalidStorageId) {
|
|
if (options & CreateOptions::OpenExisting) {
|
|
LOG(INFO) << "Opened existing storage " << id;
|
|
return id;
|
|
}
|
|
LOG(ERROR) << "Directory " << mountPoint << " is already mounted at storage " << id;
|
|
return kInvalidStorageId;
|
|
}
|
|
}
|
|
|
|
if (!(options & CreateOptions::CreateNew)) {
|
|
LOG(ERROR) << "not requirested create new storage, and it doesn't exist: " << mountPoint;
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
if (!path::isEmptyDir(mountNorm)) {
|
|
LOG(ERROR) << "Mounting over existing non-empty directory is not supported: " << mountNorm;
|
|
return kInvalidStorageId;
|
|
}
|
|
auto [mountKey, mountRoot] = makeMountDir(mIncrementalDir, mountNorm);
|
|
if (mountRoot.empty()) {
|
|
LOG(ERROR) << "Bad mount point";
|
|
return kInvalidStorageId;
|
|
}
|
|
// Make sure the code removes all crap it may create while still failing.
|
|
auto firstCleanup = [](const std::string* ptr) { IncFsMount::cleanupFilesystem(*ptr); };
|
|
auto firstCleanupOnFailure =
|
|
std::unique_ptr<std::string, decltype(firstCleanup)>(&mountRoot, firstCleanup);
|
|
|
|
auto mountTarget = path::join(mountRoot, constants().mount);
|
|
const auto backing = path::join(mountRoot, constants().backing);
|
|
if (!mkdirOrLog(backing, 0777) || !mkdirOrLog(mountTarget)) {
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
std::string metricsKey;
|
|
IncFsMount::Control control;
|
|
{
|
|
std::lock_guard l(mMountOperationLock);
|
|
IncrementalFileSystemControlParcel controlParcel;
|
|
|
|
if (auto err = rmDirContent(backing.c_str())) {
|
|
LOG(ERROR) << "Coudn't clean the backing directory " << backing << ": " << err;
|
|
return kInvalidStorageId;
|
|
}
|
|
if (!mkdirOrLog(path::join(backing, ".index"), 0777)) {
|
|
return kInvalidStorageId;
|
|
}
|
|
if (!mkdirOrLog(path::join(backing, ".incomplete"), 0777)) {
|
|
return kInvalidStorageId;
|
|
}
|
|
metricsKey = makeUniqueName(mountKey);
|
|
auto status = mVold->mountIncFs(backing, mountTarget, 0, metricsKey, &controlParcel);
|
|
if (!status.isOk()) {
|
|
LOG(ERROR) << "Vold::mountIncFs() failed: " << status.toString8();
|
|
return kInvalidStorageId;
|
|
}
|
|
if (controlParcel.cmd.get() < 0 || controlParcel.pendingReads.get() < 0 ||
|
|
controlParcel.log.get() < 0) {
|
|
LOG(ERROR) << "Vold::mountIncFs() returned invalid control parcel.";
|
|
return kInvalidStorageId;
|
|
}
|
|
int cmd = controlParcel.cmd.release().release();
|
|
int pendingReads = controlParcel.pendingReads.release().release();
|
|
int logs = controlParcel.log.release().release();
|
|
int blocksWritten =
|
|
controlParcel.blocksWritten ? controlParcel.blocksWritten->release().release() : -1;
|
|
control = mIncFs->createControl(cmd, pendingReads, logs, blocksWritten);
|
|
}
|
|
|
|
std::unique_lock l(mLock);
|
|
const auto mountIt = getStorageSlotLocked();
|
|
const auto mountId = mountIt->first;
|
|
l.unlock();
|
|
|
|
auto ifs = std::make_shared<IncFsMount>(std::move(mountRoot), std::move(metricsKey), mountId,
|
|
std::move(control), *this);
|
|
// Now it's the |ifs|'s responsibility to clean up after itself, and the only cleanup we need
|
|
// is the removal of the |ifs|.
|
|
(void)firstCleanupOnFailure.release();
|
|
|
|
auto secondCleanup = [this, &l](auto itPtr) {
|
|
if (!l.owns_lock()) {
|
|
l.lock();
|
|
}
|
|
mMounts.erase(*itPtr);
|
|
};
|
|
auto secondCleanupOnFailure =
|
|
std::unique_ptr<decltype(mountIt), decltype(secondCleanup)>(&mountIt, secondCleanup);
|
|
|
|
const auto storageIt = ifs->makeStorage(ifs->mountId);
|
|
if (storageIt == ifs->storages.end()) {
|
|
LOG(ERROR) << "Can't create a default storage directory";
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
{
|
|
metadata::Mount m;
|
|
m.mutable_storage()->set_id(ifs->mountId);
|
|
m.mutable_loader()->set_type((int)dataLoaderParams.type);
|
|
m.mutable_loader()->set_package_name(std::move(dataLoaderParams.packageName));
|
|
m.mutable_loader()->set_class_name(std::move(dataLoaderParams.className));
|
|
m.mutable_loader()->set_arguments(std::move(dataLoaderParams.arguments));
|
|
const auto metadata = m.SerializeAsString();
|
|
if (auto err =
|
|
mIncFs->makeFile(ifs->control,
|
|
path::join(ifs->root, constants().mount,
|
|
constants().infoMdName),
|
|
0777, idFromMetadata(metadata),
|
|
{.metadata = {metadata.data(), (IncFsSize)metadata.size()}})) {
|
|
LOG(ERROR) << "Saving mount metadata failed: " << -err;
|
|
return kInvalidStorageId;
|
|
}
|
|
}
|
|
|
|
const auto bk =
|
|
(options & CreateOptions::PermanentBind) ? BindKind::Permanent : BindKind::Temporary;
|
|
if (auto err = addBindMount(*ifs, storageIt->first, storageIt->second.name,
|
|
std::string(storageIt->second.name), std::move(mountNorm), bk, l);
|
|
err < 0) {
|
|
LOG(ERROR) << "Adding bind mount failed: " << -err;
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
// Done here as well, all data structures are in good state.
|
|
(void)secondCleanupOnFailure.release();
|
|
|
|
mountIt->second = std::move(ifs);
|
|
l.unlock();
|
|
|
|
LOG(INFO) << "created storage " << mountId;
|
|
return mountId;
|
|
}
|
|
|
|
StorageId IncrementalService::createLinkedStorage(std::string_view mountPoint,
|
|
StorageId linkedStorage,
|
|
IncrementalService::CreateOptions options) {
|
|
if (!isValidMountTarget(mountPoint)) {
|
|
LOG(ERROR) << "Mount point is invalid or missing";
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
std::unique_lock l(mLock);
|
|
auto ifs = getIfsLocked(linkedStorage);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "Ifs unavailable";
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
const auto mountIt = getStorageSlotLocked();
|
|
const auto storageId = mountIt->first;
|
|
const auto storageIt = ifs->makeStorage(storageId);
|
|
if (storageIt == ifs->storages.end()) {
|
|
LOG(ERROR) << "Can't create a new storage";
|
|
mMounts.erase(mountIt);
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
l.unlock();
|
|
|
|
const auto bk =
|
|
(options & CreateOptions::PermanentBind) ? BindKind::Permanent : BindKind::Temporary;
|
|
if (auto err = addBindMount(*ifs, storageIt->first, storageIt->second.name,
|
|
std::string(storageIt->second.name), path::normalize(mountPoint),
|
|
bk, l);
|
|
err < 0) {
|
|
LOG(ERROR) << "bindMount failed with error: " << err;
|
|
(void)mIncFs->unlink(ifs->control, storageIt->second.name);
|
|
ifs->storages.erase(storageIt);
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
mountIt->second = ifs;
|
|
return storageId;
|
|
}
|
|
|
|
bool IncrementalService::startLoading(StorageId storageId,
|
|
content::pm::DataLoaderParamsParcel dataLoaderParams,
|
|
DataLoaderStatusListener statusListener,
|
|
const StorageHealthCheckParams& healthCheckParams,
|
|
StorageHealthListener healthListener,
|
|
std::vector<PerUidReadTimeouts> perUidReadTimeouts) {
|
|
// Per Uid timeouts.
|
|
if (!perUidReadTimeouts.empty()) {
|
|
setUidReadTimeouts(storageId, std::move(perUidReadTimeouts));
|
|
}
|
|
|
|
IfsMountPtr ifs;
|
|
DataLoaderStubPtr dataLoaderStub;
|
|
|
|
// Re-initialize DataLoader.
|
|
{
|
|
ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
return false;
|
|
}
|
|
|
|
std::unique_lock l(ifs->lock);
|
|
dataLoaderStub = std::exchange(ifs->dataLoaderStub, nullptr);
|
|
}
|
|
|
|
if (dataLoaderStub) {
|
|
dataLoaderStub->cleanupResources();
|
|
dataLoaderStub = {};
|
|
}
|
|
|
|
{
|
|
std::unique_lock l(ifs->lock);
|
|
if (ifs->dataLoaderStub) {
|
|
LOG(INFO) << "Skipped data loader stub creation because it already exists";
|
|
return false;
|
|
}
|
|
|
|
prepareDataLoaderLocked(*ifs, std::move(dataLoaderParams), std::move(statusListener),
|
|
healthCheckParams, std::move(healthListener));
|
|
CHECK(ifs->dataLoaderStub);
|
|
dataLoaderStub = ifs->dataLoaderStub;
|
|
|
|
// Disable long read timeouts for non-system dataloaders.
|
|
// To be re-enabled after installation is complete.
|
|
ifs->setReadTimeoutsRequested(dataLoaderStub->isSystemDataLoader() &&
|
|
getAlwaysEnableReadTimeoutsForSystemDataLoaders());
|
|
applyStorageParamsLocked(*ifs);
|
|
}
|
|
|
|
if (dataLoaderStub->isSystemDataLoader() &&
|
|
!getEnforceReadLogsMaxIntervalForSystemDataLoaders()) {
|
|
// Readlogs from system dataloader (adb) can always be collected.
|
|
ifs->startLoadingTs = TimePoint::max();
|
|
} else {
|
|
// Assign time when installation wants the DL to start streaming.
|
|
const auto startLoadingTs = mClock->now();
|
|
ifs->startLoadingTs = startLoadingTs;
|
|
// Setup a callback to disable the readlogs after max interval.
|
|
addTimedJob(*mTimedQueue, storageId, getReadLogsMaxInterval(),
|
|
[this, storageId, startLoadingTs]() {
|
|
const auto ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
LOG(WARNING) << "Can't disable the readlogs, invalid storageId: "
|
|
<< storageId;
|
|
return;
|
|
}
|
|
std::unique_lock l(ifs->lock);
|
|
if (ifs->startLoadingTs != startLoadingTs) {
|
|
LOG(INFO) << "Can't disable the readlogs, timestamp mismatch (new "
|
|
"installation?): "
|
|
<< storageId;
|
|
return;
|
|
}
|
|
disableReadLogsLocked(*ifs);
|
|
});
|
|
}
|
|
|
|
return dataLoaderStub->requestStart();
|
|
}
|
|
|
|
void IncrementalService::onInstallationComplete(StorageId storage) {
|
|
IfsMountPtr ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
return;
|
|
}
|
|
|
|
// Always enable long read timeouts after installation is complete.
|
|
std::unique_lock l(ifs->lock);
|
|
ifs->setReadTimeoutsRequested(true);
|
|
applyStorageParamsLocked(*ifs);
|
|
}
|
|
|
|
IncrementalService::BindPathMap::const_iterator IncrementalService::findStorageLocked(
|
|
std::string_view path) const {
|
|
return findParentPath(mBindsByPath, path);
|
|
}
|
|
|
|
StorageId IncrementalService::findStorageId(std::string_view path) const {
|
|
std::lock_guard l(mLock);
|
|
auto it = findStorageLocked(path);
|
|
if (it == mBindsByPath.end()) {
|
|
return kInvalidStorageId;
|
|
}
|
|
return it->second->second.storage;
|
|
}
|
|
|
|
void IncrementalService::disallowReadLogs(StorageId storageId) {
|
|
const auto ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "disallowReadLogs failed, invalid storageId: " << storageId;
|
|
return;
|
|
}
|
|
|
|
std::unique_lock l(ifs->lock);
|
|
if (!ifs->readLogsAllowed()) {
|
|
return;
|
|
}
|
|
ifs->disallowReadLogs();
|
|
|
|
const auto metadata = constants().readLogsDisabledMarkerName;
|
|
if (auto err = mIncFs->makeFile(ifs->control,
|
|
path::join(ifs->root, constants().mount,
|
|
constants().readLogsDisabledMarkerName),
|
|
0777, idFromMetadata(metadata), {})) {
|
|
//{.metadata = {metadata.data(), (IncFsSize)metadata.size()}})) {
|
|
LOG(ERROR) << "Failed to make marker file for storageId: " << storageId;
|
|
return;
|
|
}
|
|
|
|
disableReadLogsLocked(*ifs);
|
|
}
|
|
|
|
int IncrementalService::setStorageParams(StorageId storageId, bool enableReadLogs) {
|
|
const auto ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "setStorageParams failed, invalid storageId: " << storageId;
|
|
return -EINVAL;
|
|
}
|
|
|
|
std::string packageName;
|
|
|
|
{
|
|
std::unique_lock l(ifs->lock);
|
|
if (!enableReadLogs) {
|
|
return disableReadLogsLocked(*ifs);
|
|
}
|
|
|
|
if (!ifs->readLogsAllowed()) {
|
|
LOG(ERROR) << "enableReadLogs failed, readlogs disallowed for storageId: " << storageId;
|
|
return -EPERM;
|
|
}
|
|
|
|
if (!ifs->dataLoaderStub) {
|
|
// This should never happen - only DL can call enableReadLogs.
|
|
LOG(ERROR) << "enableReadLogs failed: invalid state";
|
|
return -EPERM;
|
|
}
|
|
|
|
// Check installation time.
|
|
const auto now = mClock->now();
|
|
const auto startLoadingTs = ifs->startLoadingTs;
|
|
if (startLoadingTs <= now && now - startLoadingTs > getReadLogsMaxInterval()) {
|
|
LOG(ERROR)
|
|
<< "enableReadLogs failed, readlogs can't be enabled at this time, storageId: "
|
|
<< storageId;
|
|
return -EPERM;
|
|
}
|
|
|
|
packageName = ifs->dataLoaderStub->params().packageName;
|
|
ifs->setReadLogsRequested(true);
|
|
}
|
|
|
|
// Check loader usage stats permission and apop.
|
|
if (auto status =
|
|
mAppOpsManager->checkPermission(kLoaderUsageStats, kOpUsage, packageName.c_str());
|
|
!status.isOk()) {
|
|
LOG(ERROR) << " Permission: " << kLoaderUsageStats
|
|
<< " check failed: " << status.toString8();
|
|
return fromBinderStatus(status);
|
|
}
|
|
|
|
// Check multiuser permission.
|
|
if (auto status =
|
|
mAppOpsManager->checkPermission(kInteractAcrossUsers, nullptr, packageName.c_str());
|
|
!status.isOk()) {
|
|
LOG(ERROR) << " Permission: " << kInteractAcrossUsers
|
|
<< " check failed: " << status.toString8();
|
|
return fromBinderStatus(status);
|
|
}
|
|
|
|
{
|
|
std::unique_lock l(ifs->lock);
|
|
if (!ifs->readLogsRequested()) {
|
|
return 0;
|
|
}
|
|
if (auto status = applyStorageParamsLocked(*ifs); status != 0) {
|
|
return status;
|
|
}
|
|
}
|
|
|
|
registerAppOpsCallback(packageName);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int IncrementalService::disableReadLogsLocked(IncFsMount& ifs) {
|
|
ifs.setReadLogsRequested(false);
|
|
return applyStorageParamsLocked(ifs);
|
|
}
|
|
|
|
int IncrementalService::applyStorageParamsLocked(IncFsMount& ifs) {
|
|
os::incremental::IncrementalFileSystemControlParcel control;
|
|
control.cmd.reset(dup(ifs.control.cmd()));
|
|
control.pendingReads.reset(dup(ifs.control.pendingReads()));
|
|
auto logsFd = ifs.control.logs();
|
|
if (logsFd >= 0) {
|
|
control.log.reset(dup(logsFd));
|
|
}
|
|
|
|
bool enableReadLogs = ifs.readLogsRequested();
|
|
bool enableReadTimeouts = ifs.readTimeoutsRequested();
|
|
|
|
std::lock_guard l(mMountOperationLock);
|
|
auto status = mVold->setIncFsMountOptions(control, enableReadLogs, enableReadTimeouts,
|
|
ifs.metricsKey);
|
|
if (status.isOk()) {
|
|
// Store states.
|
|
ifs.setReadLogsEnabled(enableReadLogs);
|
|
ifs.setReadTimeoutsEnabled(enableReadTimeouts);
|
|
} else {
|
|
LOG(ERROR) << "applyStorageParams failed: " << status.toString8();
|
|
}
|
|
return status.isOk() ? 0 : fromBinderStatus(status);
|
|
}
|
|
|
|
void IncrementalService::deleteStorage(StorageId storageId) {
|
|
const auto ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
return;
|
|
}
|
|
deleteStorage(*ifs);
|
|
}
|
|
|
|
void IncrementalService::deleteStorage(IncrementalService::IncFsMount& ifs) {
|
|
std::unique_lock l(ifs.lock);
|
|
deleteStorageLocked(ifs, std::move(l));
|
|
}
|
|
|
|
void IncrementalService::deleteStorageLocked(IncrementalService::IncFsMount& ifs,
|
|
std::unique_lock<std::mutex>&& ifsLock) {
|
|
const auto storages = std::move(ifs.storages);
|
|
// Don't move the bind points out: Ifs's dtor will use them to unmount everything.
|
|
const auto bindPoints = ifs.bindPoints;
|
|
ifsLock.unlock();
|
|
|
|
std::lock_guard l(mLock);
|
|
for (auto&& [id, _] : storages) {
|
|
if (id != ifs.mountId) {
|
|
mMounts.erase(id);
|
|
}
|
|
}
|
|
for (auto&& [path, _] : bindPoints) {
|
|
mBindsByPath.erase(path);
|
|
}
|
|
mMounts.erase(ifs.mountId);
|
|
}
|
|
|
|
StorageId IncrementalService::openStorage(std::string_view pathInMount) {
|
|
if (!path::isAbsolute(pathInMount)) {
|
|
return kInvalidStorageId;
|
|
}
|
|
|
|
return findStorageId(path::normalize(pathInMount));
|
|
}
|
|
|
|
IncrementalService::IfsMountPtr IncrementalService::getIfs(StorageId storage) const {
|
|
std::lock_guard l(mLock);
|
|
return getIfsLocked(storage);
|
|
}
|
|
|
|
const IncrementalService::IfsMountPtr& IncrementalService::getIfsLocked(StorageId storage) const {
|
|
auto it = mMounts.find(storage);
|
|
if (it == mMounts.end()) {
|
|
static const base::NoDestructor<IfsMountPtr> kEmpty{};
|
|
return *kEmpty;
|
|
}
|
|
return it->second;
|
|
}
|
|
|
|
int IncrementalService::bind(StorageId storage, std::string_view source, std::string_view target,
|
|
BindKind kind) {
|
|
if (!isValidMountTarget(target)) {
|
|
LOG(ERROR) << __func__ << ": not a valid bind target " << target;
|
|
return -EINVAL;
|
|
}
|
|
|
|
const auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
LOG(ERROR) << __func__ << ": no ifs object for storage " << storage;
|
|
return -EINVAL;
|
|
}
|
|
|
|
std::unique_lock l(ifs->lock);
|
|
const auto storageInfo = ifs->storages.find(storage);
|
|
if (storageInfo == ifs->storages.end()) {
|
|
LOG(ERROR) << "no storage";
|
|
return -EINVAL;
|
|
}
|
|
std::string normSource = normalizePathToStorageLocked(*ifs, storageInfo, source);
|
|
if (normSource.empty()) {
|
|
LOG(ERROR) << "invalid source path";
|
|
return -EINVAL;
|
|
}
|
|
l.unlock();
|
|
std::unique_lock l2(mLock, std::defer_lock);
|
|
return addBindMount(*ifs, storage, storageInfo->second.name, std::move(normSource),
|
|
path::normalize(target), kind, l2);
|
|
}
|
|
|
|
int IncrementalService::unbind(StorageId storage, std::string_view target) {
|
|
if (!path::isAbsolute(target)) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
LOG(INFO) << "Removing bind point " << target << " for storage " << storage;
|
|
|
|
// Here we should only look up by the exact target, not by a subdirectory of any existing mount,
|
|
// otherwise there's a chance to unmount something completely unrelated
|
|
const auto norm = path::normalize(target);
|
|
std::unique_lock l(mLock);
|
|
const auto storageIt = mBindsByPath.find(norm);
|
|
if (storageIt == mBindsByPath.end() || storageIt->second->second.storage != storage) {
|
|
return -EINVAL;
|
|
}
|
|
const auto bindIt = storageIt->second;
|
|
const auto storageId = bindIt->second.storage;
|
|
const auto ifs = getIfsLocked(storageId);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "Internal error: storageId " << storageId << " for bound path " << target
|
|
<< " is missing";
|
|
return -EFAULT;
|
|
}
|
|
mBindsByPath.erase(storageIt);
|
|
l.unlock();
|
|
|
|
mVold->unmountIncFs(bindIt->first);
|
|
std::unique_lock l2(ifs->lock);
|
|
if (ifs->bindPoints.size() <= 1) {
|
|
ifs->bindPoints.clear();
|
|
deleteStorageLocked(*ifs, std::move(l2));
|
|
} else {
|
|
const std::string savedFile = std::move(bindIt->second.savedFilename);
|
|
ifs->bindPoints.erase(bindIt);
|
|
l2.unlock();
|
|
if (!savedFile.empty()) {
|
|
mIncFs->unlink(ifs->control, path::join(ifs->root, constants().mount, savedFile));
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
std::string IncrementalService::normalizePathToStorageLocked(
|
|
const IncFsMount& incfs, IncFsMount::StorageMap::const_iterator storageIt,
|
|
std::string_view path) const {
|
|
if (!path::isAbsolute(path)) {
|
|
return path::normalize(path::join(storageIt->second.name, path));
|
|
}
|
|
auto normPath = path::normalize(path);
|
|
if (path::startsWith(normPath, storageIt->second.name)) {
|
|
return normPath;
|
|
}
|
|
// not that easy: need to find if any of the bind points match
|
|
const auto bindIt = findParentPath(incfs.bindPoints, normPath);
|
|
if (bindIt == incfs.bindPoints.end()) {
|
|
return {};
|
|
}
|
|
return path::join(bindIt->second.sourceDir, path::relativize(bindIt->first, normPath));
|
|
}
|
|
|
|
std::string IncrementalService::normalizePathToStorage(const IncFsMount& ifs, StorageId storage,
|
|
std::string_view path) const {
|
|
std::unique_lock l(ifs.lock);
|
|
const auto storageInfo = ifs.storages.find(storage);
|
|
if (storageInfo == ifs.storages.end()) {
|
|
return {};
|
|
}
|
|
return normalizePathToStorageLocked(ifs, storageInfo, path);
|
|
}
|
|
|
|
int IncrementalService::makeFile(StorageId storage, std::string_view path, int mode, FileId id,
|
|
incfs::NewFileParams params, std::span<const uint8_t> data) {
|
|
const auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
return -EINVAL;
|
|
}
|
|
if (data.size() > params.size) {
|
|
LOG(ERROR) << "Bad data size - bigger than file size";
|
|
return -EINVAL;
|
|
}
|
|
if (!data.empty() && data.size() != params.size) {
|
|
// Writing a page is an irreversible operation, and it can't be updated with additional
|
|
// data later. Check that the last written page is complete, or we may break the file.
|
|
if (!isPageAligned(data.size())) {
|
|
LOG(ERROR) << "Bad data size - tried to write half a page?";
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
const std::string normPath = normalizePathToStorage(*ifs, storage, path);
|
|
if (normPath.empty()) {
|
|
LOG(ERROR) << "Internal error: storageId " << storage << " failed to normalize: " << path;
|
|
return -EINVAL;
|
|
}
|
|
if (auto err = mIncFs->makeFile(ifs->control, normPath, mode, id, params); err) {
|
|
LOG(ERROR) << "Internal error: storageId " << storage << " failed to makeFile [" << normPath
|
|
<< "]: " << err;
|
|
return err;
|
|
}
|
|
if (params.size > 0) {
|
|
if (auto err = mIncFs->reserveSpace(ifs->control, id, params.size)) {
|
|
if (err != -EOPNOTSUPP) {
|
|
LOG(ERROR) << "Failed to reserve space for a new file: " << err;
|
|
(void)mIncFs->unlink(ifs->control, normPath);
|
|
return err;
|
|
} else {
|
|
LOG(WARNING) << "Reserving space for backing file isn't supported, "
|
|
"may run out of disk later";
|
|
}
|
|
}
|
|
if (!data.empty()) {
|
|
if (auto err = setFileContent(ifs, id, path, data); err) {
|
|
(void)mIncFs->unlink(ifs->control, normPath);
|
|
return err;
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int IncrementalService::makeDir(StorageId storageId, std::string_view path, int mode) {
|
|
if (auto ifs = getIfs(storageId)) {
|
|
std::string normPath = normalizePathToStorage(*ifs, storageId, path);
|
|
if (normPath.empty()) {
|
|
return -EINVAL;
|
|
}
|
|
return mIncFs->makeDir(ifs->control, normPath, mode);
|
|
}
|
|
return -EINVAL;
|
|
}
|
|
|
|
int IncrementalService::makeDirs(StorageId storageId, std::string_view path, int mode) {
|
|
const auto ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
return -EINVAL;
|
|
}
|
|
return makeDirs(*ifs, storageId, path, mode);
|
|
}
|
|
|
|
int IncrementalService::makeDirs(const IncFsMount& ifs, StorageId storageId, std::string_view path,
|
|
int mode) {
|
|
std::string normPath = normalizePathToStorage(ifs, storageId, path);
|
|
if (normPath.empty()) {
|
|
return -EINVAL;
|
|
}
|
|
return mIncFs->makeDirs(ifs.control, normPath, mode);
|
|
}
|
|
|
|
int IncrementalService::link(StorageId sourceStorageId, std::string_view oldPath,
|
|
StorageId destStorageId, std::string_view newPath) {
|
|
std::unique_lock l(mLock);
|
|
auto ifsSrc = getIfsLocked(sourceStorageId);
|
|
if (!ifsSrc) {
|
|
return -EINVAL;
|
|
}
|
|
if (sourceStorageId != destStorageId && getIfsLocked(destStorageId) != ifsSrc) {
|
|
return -EINVAL;
|
|
}
|
|
l.unlock();
|
|
std::string normOldPath = normalizePathToStorage(*ifsSrc, sourceStorageId, oldPath);
|
|
std::string normNewPath = normalizePathToStorage(*ifsSrc, destStorageId, newPath);
|
|
if (normOldPath.empty() || normNewPath.empty()) {
|
|
LOG(ERROR) << "Invalid paths in link(): " << normOldPath << " | " << normNewPath;
|
|
return -EINVAL;
|
|
}
|
|
if (auto err = mIncFs->link(ifsSrc->control, normOldPath, normNewPath); err < 0) {
|
|
PLOG(ERROR) << "Failed to link " << oldPath << "[" << normOldPath << "]"
|
|
<< " to " << newPath << "[" << normNewPath << "]";
|
|
return err;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int IncrementalService::unlink(StorageId storage, std::string_view path) {
|
|
if (auto ifs = getIfs(storage)) {
|
|
std::string normOldPath = normalizePathToStorage(*ifs, storage, path);
|
|
return mIncFs->unlink(ifs->control, normOldPath);
|
|
}
|
|
return -EINVAL;
|
|
}
|
|
|
|
int IncrementalService::addBindMount(IncFsMount& ifs, StorageId storage,
|
|
std::string_view storageRoot, std::string&& source,
|
|
std::string&& target, BindKind kind,
|
|
std::unique_lock<std::mutex>& mainLock) {
|
|
if (!isValidMountTarget(target)) {
|
|
LOG(ERROR) << __func__ << ": invalid mount target " << target;
|
|
return -EINVAL;
|
|
}
|
|
|
|
std::string mdFileName;
|
|
std::string metadataFullPath;
|
|
if (kind != BindKind::Temporary) {
|
|
metadata::BindPoint bp;
|
|
bp.set_storage_id(storage);
|
|
bp.set_allocated_dest_path(&target);
|
|
bp.set_allocated_source_subdir(&source);
|
|
const auto metadata = bp.SerializeAsString();
|
|
bp.release_dest_path();
|
|
bp.release_source_subdir();
|
|
mdFileName = makeBindMdName();
|
|
metadataFullPath = path::join(ifs.root, constants().mount, mdFileName);
|
|
auto node = mIncFs->makeFile(ifs.control, metadataFullPath, 0444, idFromMetadata(metadata),
|
|
{.metadata = {metadata.data(), (IncFsSize)metadata.size()}});
|
|
if (node) {
|
|
LOG(ERROR) << __func__ << ": couldn't create a mount node " << mdFileName;
|
|
return int(node);
|
|
}
|
|
}
|
|
|
|
const auto res = addBindMountWithMd(ifs, storage, std::move(mdFileName), std::move(source),
|
|
std::move(target), kind, mainLock);
|
|
if (res) {
|
|
mIncFs->unlink(ifs.control, metadataFullPath);
|
|
}
|
|
return res;
|
|
}
|
|
|
|
int IncrementalService::addBindMountWithMd(IncrementalService::IncFsMount& ifs, StorageId storage,
|
|
std::string&& metadataName, std::string&& source,
|
|
std::string&& target, BindKind kind,
|
|
std::unique_lock<std::mutex>& mainLock) {
|
|
{
|
|
std::lock_guard l(mMountOperationLock);
|
|
const auto status = mVold->bindMount(source, target);
|
|
if (!status.isOk()) {
|
|
LOG(ERROR) << "Calling Vold::bindMount() failed: " << status.toString8();
|
|
return status.exceptionCode() == binder::Status::EX_SERVICE_SPECIFIC
|
|
? status.serviceSpecificErrorCode() > 0 ? -status.serviceSpecificErrorCode()
|
|
: status.serviceSpecificErrorCode() == 0
|
|
? -EFAULT
|
|
: status.serviceSpecificErrorCode()
|
|
: -EIO;
|
|
}
|
|
}
|
|
|
|
if (!mainLock.owns_lock()) {
|
|
mainLock.lock();
|
|
}
|
|
std::lock_guard l(ifs.lock);
|
|
addBindMountRecordLocked(ifs, storage, std::move(metadataName), std::move(source),
|
|
std::move(target), kind);
|
|
return 0;
|
|
}
|
|
|
|
void IncrementalService::addBindMountRecordLocked(IncFsMount& ifs, StorageId storage,
|
|
std::string&& metadataName, std::string&& source,
|
|
std::string&& target, BindKind kind) {
|
|
const auto [it, _] =
|
|
ifs.bindPoints.insert_or_assign(target,
|
|
IncFsMount::Bind{storage, std::move(metadataName),
|
|
std::move(source), kind});
|
|
mBindsByPath[std::move(target)] = it;
|
|
}
|
|
|
|
RawMetadata IncrementalService::getMetadata(StorageId storage, std::string_view path) const {
|
|
const auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
return {};
|
|
}
|
|
const auto normPath = normalizePathToStorage(*ifs, storage, path);
|
|
if (normPath.empty()) {
|
|
return {};
|
|
}
|
|
return mIncFs->getMetadata(ifs->control, normPath);
|
|
}
|
|
|
|
RawMetadata IncrementalService::getMetadata(StorageId storage, FileId node) const {
|
|
const auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
return {};
|
|
}
|
|
return mIncFs->getMetadata(ifs->control, node);
|
|
}
|
|
|
|
void IncrementalService::setUidReadTimeouts(StorageId storage,
|
|
std::vector<PerUidReadTimeouts>&& perUidReadTimeouts) {
|
|
using microseconds = std::chrono::microseconds;
|
|
using milliseconds = std::chrono::milliseconds;
|
|
|
|
auto maxPendingTimeUs = microseconds(0);
|
|
for (const auto& timeouts : perUidReadTimeouts) {
|
|
maxPendingTimeUs = std::max(maxPendingTimeUs, microseconds(timeouts.maxPendingTimeUs));
|
|
}
|
|
if (maxPendingTimeUs < Constants::minPerUidTimeout) {
|
|
LOG(ERROR) << "Skip setting read timeouts (maxPendingTime < Constants::minPerUidTimeout): "
|
|
<< duration_cast<milliseconds>(maxPendingTimeUs).count() << "ms < "
|
|
<< Constants::minPerUidTimeout.count() << "ms";
|
|
return;
|
|
}
|
|
|
|
const auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "Setting read timeouts failed: invalid storage id: " << storage;
|
|
return;
|
|
}
|
|
|
|
if (auto err = mIncFs->setUidReadTimeouts(ifs->control, perUidReadTimeouts); err < 0) {
|
|
LOG(ERROR) << "Setting read timeouts failed: " << -err;
|
|
return;
|
|
}
|
|
|
|
const auto timeout = Clock::now() + maxPendingTimeUs - Constants::perUidTimeoutOffset;
|
|
addIfsStateCallback(storage, [this, timeout](StorageId storageId, IfsState state) -> bool {
|
|
if (checkUidReadTimeouts(storageId, state, timeout)) {
|
|
return true;
|
|
}
|
|
clearUidReadTimeouts(storageId);
|
|
return false;
|
|
});
|
|
}
|
|
|
|
void IncrementalService::clearUidReadTimeouts(StorageId storage) {
|
|
const auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
return;
|
|
}
|
|
mIncFs->setUidReadTimeouts(ifs->control, {});
|
|
}
|
|
|
|
bool IncrementalService::checkUidReadTimeouts(StorageId storage, IfsState state,
|
|
Clock::time_point timeLimit) {
|
|
if (Clock::now() >= timeLimit) {
|
|
// Reached maximum timeout.
|
|
return false;
|
|
}
|
|
if (state.error) {
|
|
// Something is wrong, abort.
|
|
return false;
|
|
}
|
|
|
|
// Still loading?
|
|
if (state.fullyLoaded && !state.readLogsEnabled) {
|
|
return false;
|
|
}
|
|
|
|
const auto timeLeft = timeLimit - Clock::now();
|
|
if (timeLeft < Constants::progressUpdateInterval) {
|
|
// Don't bother.
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
std::unordered_set<std::string_view> IncrementalService::adoptMountedInstances() {
|
|
std::unordered_set<std::string_view> mountedRootNames;
|
|
mIncFs->listExistingMounts([this, &mountedRootNames](auto root, auto backingDir, auto binds) {
|
|
LOG(INFO) << "Existing mount: " << backingDir << "->" << root;
|
|
for (auto [source, target] : binds) {
|
|
LOG(INFO) << " bind: '" << source << "'->'" << target << "'";
|
|
LOG(INFO) << " " << path::join(root, source);
|
|
}
|
|
|
|
// Ensure it's a kind of a mount that's managed by IncrementalService
|
|
if (path::basename(root) != constants().mount ||
|
|
path::basename(backingDir) != constants().backing) {
|
|
return;
|
|
}
|
|
const auto expectedRoot = path::dirname(root);
|
|
if (path::dirname(backingDir) != expectedRoot) {
|
|
return;
|
|
}
|
|
if (path::dirname(expectedRoot) != mIncrementalDir) {
|
|
return;
|
|
}
|
|
if (!path::basename(expectedRoot).starts_with(constants().mountKeyPrefix)) {
|
|
return;
|
|
}
|
|
|
|
LOG(INFO) << "Looks like an IncrementalService-owned: " << expectedRoot;
|
|
|
|
// make sure we clean up the mount if it happens to be a bad one.
|
|
// Note: unmounting needs to run first, so the cleanup object is created _last_.
|
|
auto cleanupFiles = makeCleanup([&]() {
|
|
LOG(INFO) << "Failed to adopt existing mount, deleting files: " << expectedRoot;
|
|
IncFsMount::cleanupFilesystem(expectedRoot);
|
|
});
|
|
auto cleanupMounts = makeCleanup([&]() {
|
|
LOG(INFO) << "Failed to adopt existing mount, cleaning up: " << expectedRoot;
|
|
for (auto&& [_, target] : binds) {
|
|
mVold->unmountIncFs(std::string(target));
|
|
}
|
|
mVold->unmountIncFs(std::string(root));
|
|
});
|
|
|
|
auto control = mIncFs->openMount(root);
|
|
if (!control) {
|
|
LOG(INFO) << "failed to open mount " << root;
|
|
return;
|
|
}
|
|
|
|
auto mountRecord =
|
|
parseFromIncfs<metadata::Mount>(mIncFs.get(), control,
|
|
path::join(root, constants().infoMdName));
|
|
if (!mountRecord.has_loader() || !mountRecord.has_storage()) {
|
|
LOG(ERROR) << "Bad mount metadata in mount at " << expectedRoot;
|
|
return;
|
|
}
|
|
|
|
auto mountId = mountRecord.storage().id();
|
|
mNextId = std::max(mNextId, mountId + 1);
|
|
|
|
DataLoaderParamsParcel dataLoaderParams;
|
|
{
|
|
const auto& loader = mountRecord.loader();
|
|
dataLoaderParams.type = (content::pm::DataLoaderType)loader.type();
|
|
dataLoaderParams.packageName = loader.package_name();
|
|
dataLoaderParams.className = loader.class_name();
|
|
dataLoaderParams.arguments = loader.arguments();
|
|
}
|
|
|
|
// Not way to obtain a real sysfs key at this point - metrics will stop working after "soft"
|
|
// reboot.
|
|
std::string metricsKey{};
|
|
auto ifs = std::make_shared<IncFsMount>(std::string(expectedRoot), std::move(metricsKey),
|
|
mountId, std::move(control), *this);
|
|
(void)cleanupFiles.release(); // ifs will take care of that now
|
|
|
|
// Check if marker file present.
|
|
if (checkReadLogsDisabledMarker(root)) {
|
|
ifs->disallowReadLogs();
|
|
}
|
|
|
|
std::vector<std::pair<std::string, metadata::BindPoint>> permanentBindPoints;
|
|
auto d = openDir(root);
|
|
while (auto e = ::readdir(d.get())) {
|
|
if (e->d_type == DT_REG) {
|
|
auto name = std::string_view(e->d_name);
|
|
if (name.starts_with(constants().mountpointMdPrefix)) {
|
|
permanentBindPoints
|
|
.emplace_back(name,
|
|
parseFromIncfs<metadata::BindPoint>(mIncFs.get(),
|
|
ifs->control,
|
|
path::join(root,
|
|
name)));
|
|
if (permanentBindPoints.back().second.dest_path().empty() ||
|
|
permanentBindPoints.back().second.source_subdir().empty()) {
|
|
permanentBindPoints.pop_back();
|
|
mIncFs->unlink(ifs->control, path::join(root, name));
|
|
} else {
|
|
LOG(INFO) << "Permanent bind record: '"
|
|
<< permanentBindPoints.back().second.source_subdir() << "'->'"
|
|
<< permanentBindPoints.back().second.dest_path() << "'";
|
|
}
|
|
}
|
|
} else if (e->d_type == DT_DIR) {
|
|
if (e->d_name == "."sv || e->d_name == ".."sv) {
|
|
continue;
|
|
}
|
|
auto name = std::string_view(e->d_name);
|
|
if (name.starts_with(constants().storagePrefix)) {
|
|
int storageId;
|
|
const auto res =
|
|
std::from_chars(name.data() + constants().storagePrefix.size() + 1,
|
|
name.data() + name.size(), storageId);
|
|
if (res.ec != std::errc{} || *res.ptr != '_') {
|
|
LOG(WARNING) << "Ignoring storage with invalid name '" << name
|
|
<< "' for mount " << expectedRoot;
|
|
continue;
|
|
}
|
|
auto [_, inserted] = mMounts.try_emplace(storageId, ifs);
|
|
if (!inserted) {
|
|
LOG(WARNING) << "Ignoring storage with duplicate id " << storageId
|
|
<< " for mount " << expectedRoot;
|
|
continue;
|
|
}
|
|
ifs->storages.insert_or_assign(storageId,
|
|
IncFsMount::Storage{path::join(root, name)});
|
|
mNextId = std::max(mNextId, storageId + 1);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (ifs->storages.empty()) {
|
|
LOG(WARNING) << "No valid storages in mount " << root;
|
|
return;
|
|
}
|
|
|
|
// now match the mounted directories with what we expect to have in the metadata
|
|
{
|
|
std::unique_lock l(mLock, std::defer_lock);
|
|
for (auto&& [metadataFile, bindRecord] : permanentBindPoints) {
|
|
auto mountedIt = std::find_if(binds.begin(), binds.end(),
|
|
[&, bindRecord = bindRecord](auto&& bind) {
|
|
return bind.second == bindRecord.dest_path() &&
|
|
path::join(root, bind.first) ==
|
|
bindRecord.source_subdir();
|
|
});
|
|
if (mountedIt != binds.end()) {
|
|
LOG(INFO) << "Matched permanent bound " << bindRecord.source_subdir()
|
|
<< " to mount " << mountedIt->first;
|
|
addBindMountRecordLocked(*ifs, bindRecord.storage_id(), std::move(metadataFile),
|
|
std::move(*bindRecord.mutable_source_subdir()),
|
|
std::move(*bindRecord.mutable_dest_path()),
|
|
BindKind::Permanent);
|
|
if (mountedIt != binds.end() - 1) {
|
|
std::iter_swap(mountedIt, binds.end() - 1);
|
|
}
|
|
binds = binds.first(binds.size() - 1);
|
|
} else {
|
|
LOG(INFO) << "Didn't match permanent bound " << bindRecord.source_subdir()
|
|
<< ", mounting";
|
|
// doesn't exist - try mounting back
|
|
if (addBindMountWithMd(*ifs, bindRecord.storage_id(), std::move(metadataFile),
|
|
std::move(*bindRecord.mutable_source_subdir()),
|
|
std::move(*bindRecord.mutable_dest_path()),
|
|
BindKind::Permanent, l)) {
|
|
mIncFs->unlink(ifs->control, metadataFile);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// if anything stays in |binds| those are probably temporary binds; system restarted since
|
|
// they were mounted - so let's unmount them all.
|
|
for (auto&& [source, target] : binds) {
|
|
if (source.empty()) {
|
|
continue;
|
|
}
|
|
mVold->unmountIncFs(std::string(target));
|
|
}
|
|
(void)cleanupMounts.release(); // ifs now manages everything
|
|
|
|
if (ifs->bindPoints.empty()) {
|
|
LOG(WARNING) << "No valid bind points for mount " << expectedRoot;
|
|
deleteStorage(*ifs);
|
|
return;
|
|
}
|
|
|
|
prepareDataLoaderLocked(*ifs, std::move(dataLoaderParams));
|
|
CHECK(ifs->dataLoaderStub);
|
|
|
|
mountedRootNames.insert(path::basename(ifs->root));
|
|
|
|
// not locking here at all: we're still in the constructor, no other calls can happen
|
|
mMounts[ifs->mountId] = std::move(ifs);
|
|
});
|
|
|
|
return mountedRootNames;
|
|
}
|
|
|
|
void IncrementalService::mountExistingImages(
|
|
const std::unordered_set<std::string_view>& mountedRootNames) {
|
|
auto dir = openDir(mIncrementalDir);
|
|
if (!dir) {
|
|
PLOG(WARNING) << "Couldn't open the root incremental dir " << mIncrementalDir;
|
|
return;
|
|
}
|
|
while (auto entry = ::readdir(dir.get())) {
|
|
if (entry->d_type != DT_DIR) {
|
|
continue;
|
|
}
|
|
std::string_view name = entry->d_name;
|
|
if (!name.starts_with(constants().mountKeyPrefix)) {
|
|
continue;
|
|
}
|
|
if (mountedRootNames.find(name) != mountedRootNames.end()) {
|
|
continue;
|
|
}
|
|
const auto root = path::join(mIncrementalDir, name);
|
|
if (!mountExistingImage(root)) {
|
|
IncFsMount::cleanupFilesystem(root);
|
|
}
|
|
}
|
|
}
|
|
|
|
bool IncrementalService::mountExistingImage(std::string_view root) {
|
|
auto mountTarget = path::join(root, constants().mount);
|
|
const auto backing = path::join(root, constants().backing);
|
|
std::string mountKey(path::basename(path::dirname(mountTarget)));
|
|
|
|
IncrementalFileSystemControlParcel controlParcel;
|
|
auto metricsKey = makeUniqueName(mountKey);
|
|
auto status = mVold->mountIncFs(backing, mountTarget, 0, metricsKey, &controlParcel);
|
|
if (!status.isOk()) {
|
|
LOG(ERROR) << "Vold::mountIncFs() failed: " << status.toString8();
|
|
return false;
|
|
}
|
|
|
|
int cmd = controlParcel.cmd.release().release();
|
|
int pendingReads = controlParcel.pendingReads.release().release();
|
|
int logs = controlParcel.log.release().release();
|
|
int blocksWritten =
|
|
controlParcel.blocksWritten ? controlParcel.blocksWritten->release().release() : -1;
|
|
IncFsMount::Control control = mIncFs->createControl(cmd, pendingReads, logs, blocksWritten);
|
|
|
|
auto ifs = std::make_shared<IncFsMount>(std::string(root), std::move(metricsKey), -1,
|
|
std::move(control), *this);
|
|
|
|
auto mount = parseFromIncfs<metadata::Mount>(mIncFs.get(), ifs->control,
|
|
path::join(mountTarget, constants().infoMdName));
|
|
if (!mount.has_loader() || !mount.has_storage()) {
|
|
LOG(ERROR) << "Bad mount metadata in mount at " << root;
|
|
return false;
|
|
}
|
|
|
|
ifs->mountId = mount.storage().id();
|
|
mNextId = std::max(mNextId, ifs->mountId + 1);
|
|
|
|
// Check if marker file present.
|
|
if (checkReadLogsDisabledMarker(mountTarget)) {
|
|
ifs->disallowReadLogs();
|
|
}
|
|
|
|
// DataLoader params
|
|
DataLoaderParamsParcel dataLoaderParams;
|
|
{
|
|
const auto& loader = mount.loader();
|
|
dataLoaderParams.type = (content::pm::DataLoaderType)loader.type();
|
|
dataLoaderParams.packageName = loader.package_name();
|
|
dataLoaderParams.className = loader.class_name();
|
|
dataLoaderParams.arguments = loader.arguments();
|
|
}
|
|
|
|
prepareDataLoaderLocked(*ifs, std::move(dataLoaderParams));
|
|
CHECK(ifs->dataLoaderStub);
|
|
|
|
std::vector<std::pair<std::string, metadata::BindPoint>> bindPoints;
|
|
auto d = openDir(mountTarget);
|
|
while (auto e = ::readdir(d.get())) {
|
|
if (e->d_type == DT_REG) {
|
|
auto name = std::string_view(e->d_name);
|
|
if (name.starts_with(constants().mountpointMdPrefix)) {
|
|
bindPoints.emplace_back(name,
|
|
parseFromIncfs<metadata::BindPoint>(mIncFs.get(),
|
|
ifs->control,
|
|
path::join(mountTarget,
|
|
name)));
|
|
if (bindPoints.back().second.dest_path().empty() ||
|
|
bindPoints.back().second.source_subdir().empty()) {
|
|
bindPoints.pop_back();
|
|
mIncFs->unlink(ifs->control, path::join(ifs->root, constants().mount, name));
|
|
}
|
|
}
|
|
} else if (e->d_type == DT_DIR) {
|
|
if (e->d_name == "."sv || e->d_name == ".."sv) {
|
|
continue;
|
|
}
|
|
auto name = std::string_view(e->d_name);
|
|
if (name.starts_with(constants().storagePrefix)) {
|
|
int storageId;
|
|
const auto res = std::from_chars(name.data() + constants().storagePrefix.size() + 1,
|
|
name.data() + name.size(), storageId);
|
|
if (res.ec != std::errc{} || *res.ptr != '_') {
|
|
LOG(WARNING) << "Ignoring storage with invalid name '" << name << "' for mount "
|
|
<< root;
|
|
continue;
|
|
}
|
|
auto [_, inserted] = mMounts.try_emplace(storageId, ifs);
|
|
if (!inserted) {
|
|
LOG(WARNING) << "Ignoring storage with duplicate id " << storageId
|
|
<< " for mount " << root;
|
|
continue;
|
|
}
|
|
ifs->storages.insert_or_assign(storageId,
|
|
IncFsMount::Storage{
|
|
path::join(root, constants().mount, name)});
|
|
mNextId = std::max(mNextId, storageId + 1);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (ifs->storages.empty()) {
|
|
LOG(WARNING) << "No valid storages in mount " << root;
|
|
return false;
|
|
}
|
|
|
|
int bindCount = 0;
|
|
{
|
|
std::unique_lock l(mLock, std::defer_lock);
|
|
for (auto&& bp : bindPoints) {
|
|
bindCount += !addBindMountWithMd(*ifs, bp.second.storage_id(), std::move(bp.first),
|
|
std::move(*bp.second.mutable_source_subdir()),
|
|
std::move(*bp.second.mutable_dest_path()),
|
|
BindKind::Permanent, l);
|
|
}
|
|
}
|
|
|
|
if (bindCount == 0) {
|
|
LOG(WARNING) << "No valid bind points for mount " << root;
|
|
deleteStorage(*ifs);
|
|
return false;
|
|
}
|
|
|
|
// not locking here at all: we're still in the constructor, no other calls can happen
|
|
mMounts[ifs->mountId] = std::move(ifs);
|
|
return true;
|
|
}
|
|
|
|
void IncrementalService::runCmdLooper() {
|
|
constexpr auto kTimeoutMsecs = -1;
|
|
while (mRunning.load(std::memory_order_relaxed)) {
|
|
mLooper->pollAll(kTimeoutMsecs);
|
|
}
|
|
}
|
|
|
|
void IncrementalService::trimReservedSpaceV1(const IncFsMount& ifs) {
|
|
mIncFs->forEachFile(ifs.control, [this](auto&& control, auto&& fileId) {
|
|
if (mIncFs->isFileFullyLoaded(control, fileId) == incfs::LoadingState::Full) {
|
|
mIncFs->reserveSpace(control, fileId, -1);
|
|
}
|
|
return true;
|
|
});
|
|
}
|
|
|
|
void IncrementalService::prepareDataLoaderLocked(IncFsMount& ifs, DataLoaderParamsParcel&& params,
|
|
DataLoaderStatusListener&& statusListener,
|
|
const StorageHealthCheckParams& healthCheckParams,
|
|
StorageHealthListener&& healthListener) {
|
|
FileSystemControlParcel fsControlParcel;
|
|
fsControlParcel.incremental = std::make_optional<IncrementalFileSystemControlParcel>();
|
|
fsControlParcel.incremental->cmd.reset(dup(ifs.control.cmd()));
|
|
fsControlParcel.incremental->pendingReads.reset(dup(ifs.control.pendingReads()));
|
|
fsControlParcel.incremental->log.reset(dup(ifs.control.logs()));
|
|
if (ifs.control.blocksWritten() >= 0) {
|
|
fsControlParcel.incremental->blocksWritten.emplace(dup(ifs.control.blocksWritten()));
|
|
}
|
|
fsControlParcel.service = new IncrementalServiceConnector(*this, ifs.mountId);
|
|
|
|
ifs.dataLoaderStub =
|
|
new DataLoaderStub(*this, ifs.mountId, std::move(params), std::move(fsControlParcel),
|
|
std::move(statusListener), healthCheckParams,
|
|
std::move(healthListener), path::join(ifs.root, constants().mount));
|
|
|
|
// pre-v2 IncFS doesn't do automatic reserved space trimming - need to run it manually
|
|
if (!(mIncFs->features() & incfs::Features::v2)) {
|
|
addIfsStateCallback(ifs.mountId, [this](StorageId storageId, IfsState state) -> bool {
|
|
if (!state.fullyLoaded) {
|
|
return true;
|
|
}
|
|
|
|
const auto ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
return false;
|
|
}
|
|
trimReservedSpaceV1(*ifs);
|
|
return false;
|
|
});
|
|
}
|
|
|
|
addIfsStateCallback(ifs.mountId, [this](StorageId storageId, IfsState state) -> bool {
|
|
if (!state.fullyLoaded || state.readLogsEnabled) {
|
|
return true;
|
|
}
|
|
|
|
DataLoaderStubPtr dataLoaderStub;
|
|
{
|
|
const auto ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
return false;
|
|
}
|
|
|
|
std::unique_lock l(ifs->lock);
|
|
dataLoaderStub = std::exchange(ifs->dataLoaderStub, nullptr);
|
|
}
|
|
|
|
if (dataLoaderStub) {
|
|
dataLoaderStub->cleanupResources();
|
|
}
|
|
|
|
return false;
|
|
});
|
|
}
|
|
|
|
template <class Duration>
|
|
static constexpr auto castToMs(Duration d) {
|
|
return std::chrono::duration_cast<std::chrono::milliseconds>(d);
|
|
}
|
|
|
|
// Extract lib files from zip, create new files in incfs and write data to them
|
|
// Lib files should be placed next to the APK file in the following matter:
|
|
// Example:
|
|
// /path/to/base.apk
|
|
// /path/to/lib/arm/first.so
|
|
// /path/to/lib/arm/second.so
|
|
bool IncrementalService::configureNativeBinaries(StorageId storage, std::string_view apkFullPath,
|
|
std::string_view libDirRelativePath,
|
|
std::string_view abi, bool extractNativeLibs) {
|
|
auto start = Clock::now();
|
|
|
|
const auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "Invalid storage " << storage;
|
|
return false;
|
|
}
|
|
|
|
const auto targetLibPathRelativeToStorage =
|
|
path::join(path::dirname(normalizePathToStorage(*ifs, storage, apkFullPath)),
|
|
libDirRelativePath);
|
|
|
|
// First prepare target directories if they don't exist yet
|
|
if (auto res = makeDirs(*ifs, storage, targetLibPathRelativeToStorage, 0755)) {
|
|
LOG(ERROR) << "Failed to prepare target lib directory " << targetLibPathRelativeToStorage
|
|
<< " errno: " << res;
|
|
return false;
|
|
}
|
|
|
|
auto mkDirsTs = Clock::now();
|
|
ZipArchiveHandle zipFileHandle;
|
|
if (OpenArchive(path::c_str(apkFullPath), &zipFileHandle)) {
|
|
LOG(ERROR) << "Failed to open zip file at " << apkFullPath;
|
|
return false;
|
|
}
|
|
|
|
// Need a shared pointer: will be passing it into all unpacking jobs.
|
|
std::shared_ptr<ZipArchive> zipFile(zipFileHandle, [](ZipArchiveHandle h) { CloseArchive(h); });
|
|
void* cookie = nullptr;
|
|
const auto libFilePrefix = path::join(constants().libDir, abi) += "/";
|
|
if (StartIteration(zipFile.get(), &cookie, libFilePrefix, constants().libSuffix)) {
|
|
LOG(ERROR) << "Failed to start zip iteration for " << apkFullPath;
|
|
return false;
|
|
}
|
|
auto endIteration = [](void* cookie) { EndIteration(cookie); };
|
|
auto iterationCleaner = std::unique_ptr<void, decltype(endIteration)>(cookie, endIteration);
|
|
|
|
auto openZipTs = Clock::now();
|
|
|
|
auto mapFiles = (mIncFs->features() & incfs::Features::v2);
|
|
incfs::FileId sourceId;
|
|
if (mapFiles) {
|
|
sourceId = mIncFs->getFileId(ifs->control, apkFullPath);
|
|
if (!incfs::isValidFileId(sourceId)) {
|
|
LOG(WARNING) << "Error getting IncFS file ID for apk path '" << apkFullPath
|
|
<< "', mapping disabled";
|
|
mapFiles = false;
|
|
}
|
|
}
|
|
|
|
std::vector<Job> jobQueue;
|
|
ZipEntry entry;
|
|
std::string_view fileName;
|
|
while (!Next(cookie, &entry, &fileName)) {
|
|
if (fileName.empty()) {
|
|
continue;
|
|
}
|
|
|
|
const auto entryUncompressed = entry.method == kCompressStored;
|
|
const auto entryPageAligned = isPageAligned(entry.offset);
|
|
|
|
if (!extractNativeLibs) {
|
|
// ensure the file is properly aligned and unpacked
|
|
if (!entryUncompressed) {
|
|
LOG(WARNING) << "Library " << fileName << " must be uncompressed to mmap it";
|
|
return false;
|
|
}
|
|
if (!entryPageAligned) {
|
|
LOG(WARNING) << "Library " << fileName
|
|
<< " must be page-aligned to mmap it, offset = 0x" << std::hex
|
|
<< entry.offset;
|
|
return false;
|
|
}
|
|
continue;
|
|
}
|
|
|
|
auto startFileTs = Clock::now();
|
|
|
|
const auto libName = path::basename(fileName);
|
|
auto targetLibPath = path::join(targetLibPathRelativeToStorage, libName);
|
|
const auto targetLibPathAbsolute = normalizePathToStorage(*ifs, storage, targetLibPath);
|
|
// If the extract file already exists, skip
|
|
if (access(targetLibPathAbsolute.c_str(), F_OK) == 0) {
|
|
if (perfLoggingEnabled()) {
|
|
LOG(INFO) << "incfs: Native lib file already exists: " << targetLibPath
|
|
<< "; skipping extraction, spent "
|
|
<< elapsedMcs(startFileTs, Clock::now()) << "mcs";
|
|
}
|
|
continue;
|
|
}
|
|
|
|
if (mapFiles && entryUncompressed && entryPageAligned && entry.uncompressed_length > 0) {
|
|
incfs::NewMappedFileParams mappedFileParams = {
|
|
.sourceId = sourceId,
|
|
.sourceOffset = entry.offset,
|
|
.size = entry.uncompressed_length,
|
|
};
|
|
|
|
if (auto res = mIncFs->makeMappedFile(ifs->control, targetLibPathAbsolute, 0755,
|
|
mappedFileParams);
|
|
res == 0) {
|
|
if (perfLoggingEnabled()) {
|
|
auto doneTs = Clock::now();
|
|
LOG(INFO) << "incfs: Mapped " << libName << ": "
|
|
<< elapsedMcs(startFileTs, doneTs) << "mcs";
|
|
}
|
|
continue;
|
|
} else {
|
|
LOG(WARNING) << "Failed to map file for: '" << targetLibPath << "' errno: " << res
|
|
<< "; falling back to full extraction";
|
|
}
|
|
}
|
|
|
|
// Create new lib file without signature info
|
|
incfs::NewFileParams libFileParams = {
|
|
.size = entry.uncompressed_length,
|
|
.signature = {},
|
|
// Metadata of the new lib file is its relative path
|
|
.metadata = {targetLibPath.c_str(), (IncFsSize)targetLibPath.size()},
|
|
};
|
|
incfs::FileId libFileId = idFromMetadata(targetLibPath);
|
|
if (auto res = mIncFs->makeFile(ifs->control, targetLibPathAbsolute, 0755, libFileId,
|
|
libFileParams)) {
|
|
LOG(ERROR) << "Failed to make file for: " << targetLibPath << " errno: " << res;
|
|
// If one lib file fails to be created, abort others as well
|
|
return false;
|
|
}
|
|
|
|
auto makeFileTs = Clock::now();
|
|
|
|
// If it is a zero-byte file, skip data writing
|
|
if (entry.uncompressed_length == 0) {
|
|
if (perfLoggingEnabled()) {
|
|
LOG(INFO) << "incfs: Extracted " << libName
|
|
<< "(0 bytes): " << elapsedMcs(startFileTs, makeFileTs) << "mcs";
|
|
}
|
|
continue;
|
|
}
|
|
|
|
jobQueue.emplace_back([this, zipFile, entry, ifs = std::weak_ptr<IncFsMount>(ifs),
|
|
libFileId, libPath = std::move(targetLibPath),
|
|
makeFileTs]() mutable {
|
|
extractZipFile(ifs.lock(), zipFile.get(), entry, libFileId, libPath, makeFileTs);
|
|
});
|
|
|
|
if (perfLoggingEnabled()) {
|
|
auto prepareJobTs = Clock::now();
|
|
LOG(INFO) << "incfs: Processed " << libName << ": "
|
|
<< elapsedMcs(startFileTs, prepareJobTs)
|
|
<< "mcs, make file: " << elapsedMcs(startFileTs, makeFileTs)
|
|
<< " prepare job: " << elapsedMcs(makeFileTs, prepareJobTs);
|
|
}
|
|
}
|
|
|
|
auto processedTs = Clock::now();
|
|
|
|
if (!jobQueue.empty()) {
|
|
{
|
|
std::lock_guard lock(mJobMutex);
|
|
if (mRunning) {
|
|
auto& existingJobs = mJobQueue[ifs->mountId];
|
|
if (existingJobs.empty()) {
|
|
existingJobs = std::move(jobQueue);
|
|
} else {
|
|
existingJobs.insert(existingJobs.end(), std::move_iterator(jobQueue.begin()),
|
|
std::move_iterator(jobQueue.end()));
|
|
}
|
|
}
|
|
}
|
|
mJobCondition.notify_all();
|
|
}
|
|
|
|
if (perfLoggingEnabled()) {
|
|
auto end = Clock::now();
|
|
LOG(INFO) << "incfs: configureNativeBinaries complete in " << elapsedMcs(start, end)
|
|
<< "mcs, make dirs: " << elapsedMcs(start, mkDirsTs)
|
|
<< " open zip: " << elapsedMcs(mkDirsTs, openZipTs)
|
|
<< " make files: " << elapsedMcs(openZipTs, processedTs)
|
|
<< " schedule jobs: " << elapsedMcs(processedTs, end);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void IncrementalService::extractZipFile(const IfsMountPtr& ifs, ZipArchiveHandle zipFile,
|
|
ZipEntry& entry, const incfs::FileId& libFileId,
|
|
std::string_view debugLibPath,
|
|
Clock::time_point scheduledTs) {
|
|
if (!ifs) {
|
|
LOG(INFO) << "Skipping zip file " << debugLibPath << " extraction for an expired mount";
|
|
return;
|
|
}
|
|
|
|
auto startedTs = Clock::now();
|
|
|
|
// Write extracted data to new file
|
|
// NOTE: don't zero-initialize memory, it may take a while for nothing
|
|
auto libData = std::unique_ptr<uint8_t[]>(new uint8_t[entry.uncompressed_length]);
|
|
if (ExtractToMemory(zipFile, &entry, libData.get(), entry.uncompressed_length)) {
|
|
LOG(ERROR) << "Failed to extract native lib zip entry: " << path::basename(debugLibPath);
|
|
return;
|
|
}
|
|
|
|
auto extractFileTs = Clock::now();
|
|
|
|
if (setFileContent(ifs, libFileId, debugLibPath,
|
|
std::span(libData.get(), entry.uncompressed_length))) {
|
|
return;
|
|
}
|
|
|
|
if (perfLoggingEnabled()) {
|
|
auto endFileTs = Clock::now();
|
|
LOG(INFO) << "incfs: Extracted " << path::basename(debugLibPath) << "("
|
|
<< entry.compressed_length << " -> " << entry.uncompressed_length
|
|
<< " bytes): " << elapsedMcs(startedTs, endFileTs)
|
|
<< "mcs, scheduling delay: " << elapsedMcs(scheduledTs, startedTs)
|
|
<< " extract: " << elapsedMcs(startedTs, extractFileTs)
|
|
<< " open/prepare/write: " << elapsedMcs(extractFileTs, endFileTs);
|
|
}
|
|
}
|
|
|
|
bool IncrementalService::waitForNativeBinariesExtraction(StorageId storage) {
|
|
struct WaitPrinter {
|
|
const Clock::time_point startTs = Clock::now();
|
|
~WaitPrinter() noexcept {
|
|
if (perfLoggingEnabled()) {
|
|
const auto endTs = Clock::now();
|
|
LOG(INFO) << "incfs: waitForNativeBinariesExtraction() complete in "
|
|
<< elapsedMcs(startTs, endTs) << "mcs";
|
|
}
|
|
}
|
|
} waitPrinter;
|
|
|
|
MountId mount;
|
|
{
|
|
auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
return true;
|
|
}
|
|
mount = ifs->mountId;
|
|
}
|
|
|
|
std::unique_lock lock(mJobMutex);
|
|
mJobCondition.wait(lock, [this, mount] {
|
|
return !mRunning ||
|
|
(mPendingJobsMount != mount && mJobQueue.find(mount) == mJobQueue.end());
|
|
});
|
|
return mRunning;
|
|
}
|
|
|
|
int IncrementalService::setFileContent(const IfsMountPtr& ifs, const incfs::FileId& fileId,
|
|
std::string_view debugFilePath,
|
|
std::span<const uint8_t> data) const {
|
|
auto startTs = Clock::now();
|
|
|
|
const auto writeFd = mIncFs->openForSpecialOps(ifs->control, fileId);
|
|
if (!writeFd.ok()) {
|
|
LOG(ERROR) << "Failed to open write fd for: " << debugFilePath
|
|
<< " errno: " << writeFd.get();
|
|
return writeFd.get();
|
|
}
|
|
|
|
const auto dataLength = data.size();
|
|
|
|
auto openFileTs = Clock::now();
|
|
const int numBlocks = (data.size() + constants().blockSize - 1) / constants().blockSize;
|
|
std::vector<IncFsDataBlock> instructions(numBlocks);
|
|
for (int i = 0; i < numBlocks; i++) {
|
|
const auto blockSize = std::min<long>(constants().blockSize, data.size());
|
|
instructions[i] = IncFsDataBlock{
|
|
.fileFd = writeFd.get(),
|
|
.pageIndex = static_cast<IncFsBlockIndex>(i),
|
|
.compression = INCFS_COMPRESSION_KIND_NONE,
|
|
.kind = INCFS_BLOCK_KIND_DATA,
|
|
.dataSize = static_cast<uint32_t>(blockSize),
|
|
.data = reinterpret_cast<const char*>(data.data()),
|
|
};
|
|
data = data.subspan(blockSize);
|
|
}
|
|
auto prepareInstsTs = Clock::now();
|
|
|
|
size_t res = mIncFs->writeBlocks(instructions);
|
|
if (res != instructions.size()) {
|
|
LOG(ERROR) << "Failed to write data into: " << debugFilePath;
|
|
return res;
|
|
}
|
|
|
|
if (perfLoggingEnabled()) {
|
|
auto endTs = Clock::now();
|
|
LOG(INFO) << "incfs: Set file content " << debugFilePath << "(" << dataLength
|
|
<< " bytes): " << elapsedMcs(startTs, endTs)
|
|
<< "mcs, open: " << elapsedMcs(startTs, openFileTs)
|
|
<< " prepare: " << elapsedMcs(openFileTs, prepareInstsTs)
|
|
<< " write: " << elapsedMcs(prepareInstsTs, endTs);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
incfs::LoadingState IncrementalService::isFileFullyLoaded(StorageId storage,
|
|
std::string_view filePath) const {
|
|
std::unique_lock l(mLock);
|
|
const auto ifs = getIfsLocked(storage);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "isFileFullyLoaded failed, invalid storageId: " << storage;
|
|
return incfs::LoadingState(-EINVAL);
|
|
}
|
|
const auto storageInfo = ifs->storages.find(storage);
|
|
if (storageInfo == ifs->storages.end()) {
|
|
LOG(ERROR) << "isFileFullyLoaded failed, no storage: " << storage;
|
|
return incfs::LoadingState(-EINVAL);
|
|
}
|
|
l.unlock();
|
|
return mIncFs->isFileFullyLoaded(ifs->control, filePath);
|
|
}
|
|
|
|
incfs::LoadingState IncrementalService::isMountFullyLoaded(StorageId storage) const {
|
|
const auto ifs = getIfs(storage);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "isMountFullyLoaded failed, invalid storageId: " << storage;
|
|
return incfs::LoadingState(-EINVAL);
|
|
}
|
|
return mIncFs->isEverythingFullyLoaded(ifs->control);
|
|
}
|
|
|
|
IncrementalService::LoadingProgress IncrementalService::getLoadingProgress(
|
|
StorageId storage) const {
|
|
std::unique_lock l(mLock);
|
|
const auto ifs = getIfsLocked(storage);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "getLoadingProgress failed, invalid storageId: " << storage;
|
|
return {-EINVAL, -EINVAL};
|
|
}
|
|
const auto storageInfo = ifs->storages.find(storage);
|
|
if (storageInfo == ifs->storages.end()) {
|
|
LOG(ERROR) << "getLoadingProgress failed, no storage: " << storage;
|
|
return {-EINVAL, -EINVAL};
|
|
}
|
|
l.unlock();
|
|
return getLoadingProgressFromPath(*ifs, storageInfo->second.name);
|
|
}
|
|
|
|
IncrementalService::LoadingProgress IncrementalService::getLoadingProgressFromPath(
|
|
const IncFsMount& ifs, std::string_view storagePath) const {
|
|
ssize_t totalBlocks = 0, filledBlocks = 0, error = 0;
|
|
mFs->listFilesRecursive(storagePath, [&, this](auto filePath) {
|
|
const auto [filledBlocksCount, totalBlocksCount] =
|
|
mIncFs->countFilledBlocks(ifs.control, filePath);
|
|
if (filledBlocksCount == -EOPNOTSUPP || filledBlocksCount == -ENOTSUP ||
|
|
filledBlocksCount == -ENOENT) {
|
|
// a kind of a file that's not really being loaded, e.g. a mapped range
|
|
// an older IncFS used to return ENOENT in this case, so handle it the same way
|
|
return true;
|
|
}
|
|
if (filledBlocksCount < 0) {
|
|
LOG(ERROR) << "getLoadingProgress failed to get filled blocks count for: " << filePath
|
|
<< ", errno: " << filledBlocksCount;
|
|
error = filledBlocksCount;
|
|
return false;
|
|
}
|
|
totalBlocks += totalBlocksCount;
|
|
filledBlocks += filledBlocksCount;
|
|
return true;
|
|
});
|
|
|
|
return error ? LoadingProgress{error, error} : LoadingProgress{filledBlocks, totalBlocks};
|
|
}
|
|
|
|
bool IncrementalService::updateLoadingProgress(StorageId storage,
|
|
StorageLoadingProgressListener&& progressListener) {
|
|
const auto progress = getLoadingProgress(storage);
|
|
if (progress.isError()) {
|
|
// Failed to get progress from incfs, abort.
|
|
return false;
|
|
}
|
|
progressListener->onStorageLoadingProgressChanged(storage, progress.getProgress());
|
|
if (progress.fullyLoaded()) {
|
|
// Stop updating progress once it is fully loaded
|
|
return true;
|
|
}
|
|
addTimedJob(*mProgressUpdateJobQueue, storage,
|
|
Constants::progressUpdateInterval /* repeat after 1s */,
|
|
[storage, progressListener = std::move(progressListener), this]() mutable {
|
|
updateLoadingProgress(storage, std::move(progressListener));
|
|
});
|
|
return true;
|
|
}
|
|
|
|
bool IncrementalService::registerLoadingProgressListener(
|
|
StorageId storage, StorageLoadingProgressListener progressListener) {
|
|
return updateLoadingProgress(storage, std::move(progressListener));
|
|
}
|
|
|
|
bool IncrementalService::unregisterLoadingProgressListener(StorageId storage) {
|
|
return removeTimedJobs(*mProgressUpdateJobQueue, storage);
|
|
}
|
|
|
|
bool IncrementalService::perfLoggingEnabled() {
|
|
static const bool enabled = base::GetBoolProperty("incremental.perflogging", false);
|
|
return enabled;
|
|
}
|
|
|
|
void IncrementalService::runJobProcessing() {
|
|
for (;;) {
|
|
std::unique_lock lock(mJobMutex);
|
|
mJobCondition.wait(lock, [this]() { return !mRunning || !mJobQueue.empty(); });
|
|
if (!mRunning) {
|
|
return;
|
|
}
|
|
|
|
auto it = mJobQueue.begin();
|
|
mPendingJobsMount = it->first;
|
|
auto queue = std::move(it->second);
|
|
mJobQueue.erase(it);
|
|
lock.unlock();
|
|
|
|
for (auto&& job : queue) {
|
|
job();
|
|
}
|
|
|
|
lock.lock();
|
|
mPendingJobsMount = kInvalidStorageId;
|
|
lock.unlock();
|
|
mJobCondition.notify_all();
|
|
}
|
|
}
|
|
|
|
void IncrementalService::registerAppOpsCallback(const std::string& packageName) {
|
|
sp<IAppOpsCallback> listener;
|
|
{
|
|
std::unique_lock lock{mCallbacksLock};
|
|
auto& cb = mCallbackRegistered[packageName];
|
|
if (cb) {
|
|
return;
|
|
}
|
|
cb = new AppOpsListener(*this, packageName);
|
|
listener = cb;
|
|
}
|
|
|
|
mAppOpsManager->startWatchingMode(AppOpsManager::OP_GET_USAGE_STATS,
|
|
String16(packageName.c_str()), listener);
|
|
}
|
|
|
|
bool IncrementalService::unregisterAppOpsCallback(const std::string& packageName) {
|
|
sp<IAppOpsCallback> listener;
|
|
{
|
|
std::unique_lock lock{mCallbacksLock};
|
|
auto found = mCallbackRegistered.find(packageName);
|
|
if (found == mCallbackRegistered.end()) {
|
|
return false;
|
|
}
|
|
listener = found->second;
|
|
mCallbackRegistered.erase(found);
|
|
}
|
|
|
|
mAppOpsManager->stopWatchingMode(listener);
|
|
return true;
|
|
}
|
|
|
|
void IncrementalService::onAppOpChanged(const std::string& packageName) {
|
|
if (!unregisterAppOpsCallback(packageName)) {
|
|
return;
|
|
}
|
|
|
|
std::vector<IfsMountPtr> affected;
|
|
{
|
|
std::lock_guard l(mLock);
|
|
affected.reserve(mMounts.size());
|
|
for (auto&& [id, ifs] : mMounts) {
|
|
std::unique_lock ll(ifs->lock);
|
|
if (ifs->mountId == id && ifs->dataLoaderStub &&
|
|
ifs->dataLoaderStub->params().packageName == packageName) {
|
|
affected.push_back(ifs);
|
|
}
|
|
}
|
|
}
|
|
for (auto&& ifs : affected) {
|
|
std::unique_lock ll(ifs->lock);
|
|
disableReadLogsLocked(*ifs);
|
|
}
|
|
}
|
|
|
|
bool IncrementalService::addTimedJob(TimedQueueWrapper& timedQueue, MountId id, Milliseconds after,
|
|
Job what) {
|
|
if (id == kInvalidStorageId) {
|
|
return false;
|
|
}
|
|
timedQueue.addJob(id, after, std::move(what));
|
|
return true;
|
|
}
|
|
|
|
bool IncrementalService::removeTimedJobs(TimedQueueWrapper& timedQueue, MountId id) {
|
|
if (id == kInvalidStorageId) {
|
|
return false;
|
|
}
|
|
timedQueue.removeJobs(id);
|
|
return true;
|
|
}
|
|
|
|
void IncrementalService::addIfsStateCallback(StorageId storageId, IfsStateCallback callback) {
|
|
bool wasEmpty;
|
|
{
|
|
std::lock_guard l(mIfsStateCallbacksLock);
|
|
wasEmpty = mIfsStateCallbacks.empty();
|
|
mIfsStateCallbacks[storageId].emplace_back(std::move(callback));
|
|
}
|
|
if (wasEmpty) {
|
|
addTimedJob(*mTimedQueue, kAllStoragesId, Constants::progressUpdateInterval,
|
|
[this]() { processIfsStateCallbacks(); });
|
|
}
|
|
}
|
|
|
|
void IncrementalService::processIfsStateCallbacks() {
|
|
StorageId storageId = kInvalidStorageId;
|
|
std::vector<IfsStateCallback> local;
|
|
while (true) {
|
|
{
|
|
std::lock_guard l(mIfsStateCallbacksLock);
|
|
if (mIfsStateCallbacks.empty()) {
|
|
return;
|
|
}
|
|
IfsStateCallbacks::iterator it;
|
|
if (storageId == kInvalidStorageId) {
|
|
// First entry, initialize the |it|.
|
|
it = mIfsStateCallbacks.begin();
|
|
} else {
|
|
// Subsequent entries, update the |storageId|, and shift to the new one (not that
|
|
// it guarantees much about updated items, but at least the loop will finish).
|
|
it = mIfsStateCallbacks.lower_bound(storageId);
|
|
if (it == mIfsStateCallbacks.end()) {
|
|
// Nothing else left, too bad.
|
|
break;
|
|
}
|
|
if (it->first != storageId) {
|
|
local.clear(); // Was removed during processing, forget the old callbacks.
|
|
} else {
|
|
// Put the 'surviving' callbacks back into the map and advance the position.
|
|
auto& callbacks = it->second;
|
|
if (callbacks.empty()) {
|
|
std::swap(callbacks, local);
|
|
} else {
|
|
callbacks.insert(callbacks.end(), std::move_iterator(local.begin()),
|
|
std::move_iterator(local.end()));
|
|
local.clear();
|
|
}
|
|
if (callbacks.empty()) {
|
|
it = mIfsStateCallbacks.erase(it);
|
|
if (mIfsStateCallbacks.empty()) {
|
|
return;
|
|
}
|
|
} else {
|
|
++it;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (it == mIfsStateCallbacks.end()) {
|
|
break;
|
|
}
|
|
|
|
storageId = it->first;
|
|
auto& callbacks = it->second;
|
|
if (callbacks.empty()) {
|
|
// Invalid case, one extra lookup should be ok.
|
|
continue;
|
|
}
|
|
std::swap(callbacks, local);
|
|
}
|
|
|
|
processIfsStateCallbacks(storageId, local);
|
|
}
|
|
|
|
addTimedJob(*mTimedQueue, kAllStoragesId, Constants::progressUpdateInterval,
|
|
[this]() { processIfsStateCallbacks(); });
|
|
}
|
|
|
|
void IncrementalService::processIfsStateCallbacks(StorageId storageId,
|
|
std::vector<IfsStateCallback>& callbacks) {
|
|
const auto state = isMountFullyLoaded(storageId);
|
|
IfsState storageState = {};
|
|
storageState.error = int(state) < 0;
|
|
storageState.fullyLoaded = state == incfs::LoadingState::Full;
|
|
if (storageState.fullyLoaded) {
|
|
const auto ifs = getIfs(storageId);
|
|
storageState.readLogsEnabled = ifs && ifs->readLogsEnabled();
|
|
}
|
|
|
|
for (auto cur = callbacks.begin(); cur != callbacks.end();) {
|
|
if ((*cur)(storageId, storageState)) {
|
|
++cur;
|
|
} else {
|
|
cur = callbacks.erase(cur);
|
|
}
|
|
}
|
|
}
|
|
|
|
void IncrementalService::removeIfsStateCallbacks(StorageId storageId) {
|
|
std::lock_guard l(mIfsStateCallbacksLock);
|
|
mIfsStateCallbacks.erase(storageId);
|
|
}
|
|
|
|
void IncrementalService::getMetrics(StorageId storageId, android::os::PersistableBundle* result) {
|
|
const auto ifs = getIfs(storageId);
|
|
if (!ifs) {
|
|
LOG(ERROR) << "getMetrics failed, invalid storageId: " << storageId;
|
|
return;
|
|
}
|
|
const auto& kMetricsReadLogsEnabled =
|
|
os::incremental::BnIncrementalService::METRICS_READ_LOGS_ENABLED();
|
|
result->putBoolean(String16(kMetricsReadLogsEnabled.c_str()), ifs->readLogsEnabled() != 0);
|
|
const auto incfsMetrics = mIncFs->getMetrics(ifs->metricsKey);
|
|
if (incfsMetrics) {
|
|
const auto& kMetricsTotalDelayedReads =
|
|
os::incremental::BnIncrementalService::METRICS_TOTAL_DELAYED_READS();
|
|
const auto totalDelayedReads =
|
|
incfsMetrics->readsDelayedMin + incfsMetrics->readsDelayedPending;
|
|
result->putInt(String16(kMetricsTotalDelayedReads.c_str()), totalDelayedReads);
|
|
const auto& kMetricsTotalFailedReads =
|
|
os::incremental::BnIncrementalService::METRICS_TOTAL_FAILED_READS();
|
|
const auto totalFailedReads = incfsMetrics->readsFailedTimedOut +
|
|
incfsMetrics->readsFailedHashVerification + incfsMetrics->readsFailedOther;
|
|
result->putInt(String16(kMetricsTotalFailedReads.c_str()), totalFailedReads);
|
|
const auto& kMetricsTotalDelayedReadsMillis =
|
|
os::incremental::BnIncrementalService::METRICS_TOTAL_DELAYED_READS_MILLIS();
|
|
const int64_t totalDelayedReadsMillis =
|
|
(incfsMetrics->readsDelayedMinUs + incfsMetrics->readsDelayedPendingUs) / 1000;
|
|
result->putLong(String16(kMetricsTotalDelayedReadsMillis.c_str()), totalDelayedReadsMillis);
|
|
}
|
|
const auto lastReadError = mIncFs->getLastReadError(ifs->control);
|
|
if (lastReadError && lastReadError->timestampUs != 0) {
|
|
const auto& kMetricsMillisSinceLastReadError =
|
|
os::incremental::BnIncrementalService::METRICS_MILLIS_SINCE_LAST_READ_ERROR();
|
|
result->putLong(String16(kMetricsMillisSinceLastReadError.c_str()),
|
|
(int64_t)elapsedUsSinceMonoTs(lastReadError->timestampUs) / 1000);
|
|
const auto& kMetricsLastReadErrorNo =
|
|
os::incremental::BnIncrementalService::METRICS_LAST_READ_ERROR_NUMBER();
|
|
result->putInt(String16(kMetricsLastReadErrorNo.c_str()), lastReadError->errorNo);
|
|
const auto& kMetricsLastReadUid =
|
|
os::incremental::BnIncrementalService::METRICS_LAST_READ_ERROR_UID();
|
|
result->putInt(String16(kMetricsLastReadUid.c_str()), lastReadError->uid);
|
|
}
|
|
std::unique_lock l(ifs->lock);
|
|
if (!ifs->dataLoaderStub) {
|
|
return;
|
|
}
|
|
ifs->dataLoaderStub->getMetrics(result);
|
|
}
|
|
|
|
IncrementalService::DataLoaderStub::DataLoaderStub(
|
|
IncrementalService& service, MountId id, DataLoaderParamsParcel&& params,
|
|
FileSystemControlParcel&& control, DataLoaderStatusListener&& statusListener,
|
|
const StorageHealthCheckParams& healthCheckParams, StorageHealthListener&& healthListener,
|
|
std::string&& healthPath)
|
|
: mService(service),
|
|
mId(id),
|
|
mParams(std::move(params)),
|
|
mControl(std::move(control)),
|
|
mStatusListener(std::move(statusListener)),
|
|
mHealthListener(std::move(healthListener)),
|
|
mHealthPath(std::move(healthPath)),
|
|
mHealthCheckParams(healthCheckParams) {
|
|
if (mHealthListener && !isHealthParamsValid()) {
|
|
mHealthListener = {};
|
|
}
|
|
if (!mHealthListener) {
|
|
// Disable advanced health check statuses.
|
|
mHealthCheckParams.blockedTimeoutMs = -1;
|
|
}
|
|
updateHealthStatus();
|
|
}
|
|
|
|
IncrementalService::DataLoaderStub::~DataLoaderStub() {
|
|
if (isValid()) {
|
|
cleanupResources();
|
|
}
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::cleanupResources() {
|
|
auto now = Clock::now();
|
|
{
|
|
std::unique_lock lock(mMutex);
|
|
mHealthPath.clear();
|
|
unregisterFromPendingReads();
|
|
resetHealthControl();
|
|
mService.removeTimedJobs(*mService.mTimedQueue, mId);
|
|
}
|
|
mService.removeIfsStateCallbacks(mId);
|
|
|
|
requestDestroy();
|
|
|
|
{
|
|
std::unique_lock lock(mMutex);
|
|
mParams = {};
|
|
mControl = {};
|
|
mHealthControl = {};
|
|
mHealthListener = {};
|
|
mStatusCondition.wait_until(lock, now + Constants::destroyTimeout, [this] {
|
|
return mCurrentStatus == IDataLoaderStatusListener::DATA_LOADER_DESTROYED;
|
|
});
|
|
mStatusListener = {};
|
|
mId = kInvalidStorageId;
|
|
}
|
|
}
|
|
|
|
sp<content::pm::IDataLoader> IncrementalService::DataLoaderStub::getDataLoader() {
|
|
sp<IDataLoader> dataloader;
|
|
auto status = mService.mDataLoaderManager->getDataLoader(id(), &dataloader);
|
|
if (!status.isOk()) {
|
|
LOG(ERROR) << "Failed to get dataloader: " << status.toString8();
|
|
return {};
|
|
}
|
|
if (!dataloader) {
|
|
LOG(ERROR) << "DataLoader is null: " << status.toString8();
|
|
return {};
|
|
}
|
|
return dataloader;
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::isSystemDataLoader() const {
|
|
return (params().packageName == Constants::systemPackage);
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::requestCreate() {
|
|
return setTargetStatus(IDataLoaderStatusListener::DATA_LOADER_CREATED);
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::requestStart() {
|
|
return setTargetStatus(IDataLoaderStatusListener::DATA_LOADER_STARTED);
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::requestDestroy() {
|
|
return setTargetStatus(IDataLoaderStatusListener::DATA_LOADER_DESTROYED);
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::setTargetStatus(int newStatus) {
|
|
{
|
|
std::unique_lock lock(mMutex);
|
|
setTargetStatusLocked(newStatus);
|
|
}
|
|
return fsmStep();
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::setTargetStatusLocked(int status) {
|
|
auto oldStatus = mTargetStatus;
|
|
mTargetStatus = status;
|
|
mTargetStatusTs = Clock::now();
|
|
LOG(DEBUG) << "Target status update for DataLoader " << id() << ": " << oldStatus << " -> "
|
|
<< status << " (current " << mCurrentStatus << ")";
|
|
}
|
|
|
|
std::optional<Milliseconds> IncrementalService::DataLoaderStub::needToBind() {
|
|
std::unique_lock lock(mMutex);
|
|
|
|
const auto now = mService.mClock->now();
|
|
const bool healthy = (mPreviousBindDelay == 0ms);
|
|
|
|
if (mCurrentStatus == IDataLoaderStatusListener::DATA_LOADER_BINDING &&
|
|
now - mCurrentStatusTs <= Constants::bindingTimeout) {
|
|
LOG(INFO) << "Binding still in progress. "
|
|
<< (healthy ? "The DL is healthy/freshly bound, ok to retry for a few times."
|
|
: "Already unhealthy, don't do anything.")
|
|
<< " for storage " << mId;
|
|
// Binding still in progress.
|
|
if (!healthy) {
|
|
// Already unhealthy, don't do anything.
|
|
return {};
|
|
}
|
|
// The DL is healthy/freshly bound, ok to retry for a few times.
|
|
if (now - mPreviousBindTs <= Constants::bindGracePeriod) {
|
|
// Still within grace period.
|
|
if (now - mCurrentStatusTs >= Constants::bindRetryInterval) {
|
|
// Retry interval passed, retrying.
|
|
mCurrentStatusTs = now;
|
|
mPreviousBindDelay = 0ms;
|
|
return 0ms;
|
|
}
|
|
return {};
|
|
}
|
|
// fallthrough, mark as unhealthy, and retry with delay
|
|
}
|
|
|
|
const auto previousBindTs = mPreviousBindTs;
|
|
mPreviousBindTs = now;
|
|
|
|
const auto nonCrashingInterval =
|
|
std::max(castToMs(now - previousBindTs - mPreviousBindDelay), 100ms);
|
|
if (previousBindTs.time_since_epoch() == Clock::duration::zero() ||
|
|
nonCrashingInterval > Constants::healthyDataLoaderUptime) {
|
|
mPreviousBindDelay = 0ms;
|
|
return 0ms;
|
|
}
|
|
|
|
constexpr auto minBindDelayMs = castToMs(Constants::minBindDelay);
|
|
constexpr auto maxBindDelayMs = castToMs(Constants::maxBindDelay);
|
|
|
|
const auto bindDelayMs =
|
|
std::min(std::max(mPreviousBindDelay * Constants::bindDelayMultiplier, minBindDelayMs),
|
|
maxBindDelayMs)
|
|
.count();
|
|
const auto bindDelayJitterRangeMs = bindDelayMs / Constants::bindDelayJitterDivider;
|
|
// rand() is enough, not worth maintaining a full-blown <rand> object for delay jitter
|
|
const auto bindDelayJitterMs = rand() % (bindDelayJitterRangeMs * 2) - // NOLINT
|
|
bindDelayJitterRangeMs;
|
|
mPreviousBindDelay = std::chrono::milliseconds(bindDelayMs + bindDelayJitterMs);
|
|
return mPreviousBindDelay;
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::bind() {
|
|
const auto maybeBindDelay = needToBind();
|
|
if (!maybeBindDelay) {
|
|
LOG(DEBUG) << "Skipping bind to " << mParams.packageName << " because of pending bind.";
|
|
return true;
|
|
}
|
|
const auto bindDelay = *maybeBindDelay;
|
|
if (bindDelay > 1s) {
|
|
LOG(INFO) << "Delaying bind to " << mParams.packageName << " by "
|
|
<< bindDelay.count() / 1000 << "s"
|
|
<< " for storage " << mId;
|
|
}
|
|
|
|
bool result = false;
|
|
auto status = mService.mDataLoaderManager->bindToDataLoader(id(), mParams, bindDelay.count(),
|
|
this, &result);
|
|
if (!status.isOk() || !result) {
|
|
const bool healthy = (bindDelay == 0ms);
|
|
LOG(ERROR) << "Failed to bind a data loader for mount " << id()
|
|
<< (healthy ? ", retrying." : "");
|
|
|
|
// Internal error, retry for healthy/new DLs.
|
|
// Let needToBind migrate it to unhealthy after too many retries.
|
|
if (healthy) {
|
|
if (mService.addTimedJob(*mService.mTimedQueue, id(), Constants::bindRetryInterval,
|
|
[this]() { fsmStep(); })) {
|
|
// Mark as binding so that we know it's not the DL's fault.
|
|
setCurrentStatus(IDataLoaderStatusListener::DATA_LOADER_BINDING);
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::create() {
|
|
auto dataloader = getDataLoader();
|
|
if (!dataloader) {
|
|
return false;
|
|
}
|
|
auto status = dataloader->create(id(), mParams, mControl, this);
|
|
if (!status.isOk()) {
|
|
LOG(ERROR) << "Failed to create DataLoader: " << status.toString8();
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::start() {
|
|
auto dataloader = getDataLoader();
|
|
if (!dataloader) {
|
|
return false;
|
|
}
|
|
auto status = dataloader->start(id());
|
|
if (!status.isOk()) {
|
|
LOG(ERROR) << "Failed to start DataLoader: " << status.toString8();
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::destroy() {
|
|
return mService.mDataLoaderManager->unbindFromDataLoader(id()).isOk();
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::fsmStep() {
|
|
if (!isValid()) {
|
|
return false;
|
|
}
|
|
|
|
int currentStatus;
|
|
int targetStatus;
|
|
{
|
|
std::unique_lock lock(mMutex);
|
|
currentStatus = mCurrentStatus;
|
|
targetStatus = mTargetStatus;
|
|
}
|
|
|
|
LOG(DEBUG) << "fsmStep: " << id() << ": " << currentStatus << " -> " << targetStatus;
|
|
|
|
if (currentStatus == targetStatus) {
|
|
return true;
|
|
}
|
|
|
|
switch (targetStatus) {
|
|
case IDataLoaderStatusListener::DATA_LOADER_DESTROYED: {
|
|
switch (currentStatus) {
|
|
case IDataLoaderStatusListener::DATA_LOADER_UNAVAILABLE:
|
|
case IDataLoaderStatusListener::DATA_LOADER_UNRECOVERABLE:
|
|
destroy();
|
|
// DataLoader is broken, just assume it's destroyed.
|
|
compareAndSetCurrentStatus(currentStatus,
|
|
IDataLoaderStatusListener::DATA_LOADER_DESTROYED);
|
|
return true;
|
|
case IDataLoaderStatusListener::DATA_LOADER_BINDING:
|
|
compareAndSetCurrentStatus(currentStatus,
|
|
IDataLoaderStatusListener::DATA_LOADER_DESTROYED);
|
|
return true;
|
|
default:
|
|
return destroy();
|
|
}
|
|
break;
|
|
}
|
|
case IDataLoaderStatusListener::DATA_LOADER_STARTED: {
|
|
switch (currentStatus) {
|
|
case IDataLoaderStatusListener::DATA_LOADER_CREATED:
|
|
case IDataLoaderStatusListener::DATA_LOADER_STOPPED:
|
|
return start();
|
|
}
|
|
[[fallthrough]];
|
|
}
|
|
case IDataLoaderStatusListener::DATA_LOADER_CREATED:
|
|
switch (currentStatus) {
|
|
case IDataLoaderStatusListener::DATA_LOADER_UNAVAILABLE:
|
|
case IDataLoaderStatusListener::DATA_LOADER_UNRECOVERABLE:
|
|
// Before binding need to make sure we are unbound.
|
|
// Otherwise we'll get stuck binding.
|
|
destroy();
|
|
// DataLoader is broken, just assume it's destroyed.
|
|
compareAndSetCurrentStatus(currentStatus,
|
|
IDataLoaderStatusListener::DATA_LOADER_DESTROYED);
|
|
return true;
|
|
case IDataLoaderStatusListener::DATA_LOADER_DESTROYED:
|
|
case IDataLoaderStatusListener::DATA_LOADER_BINDING:
|
|
return bind();
|
|
case IDataLoaderStatusListener::DATA_LOADER_BOUND:
|
|
return create();
|
|
}
|
|
break;
|
|
default:
|
|
LOG(ERROR) << "Invalid target status: " << targetStatus
|
|
<< ", current status: " << currentStatus;
|
|
break;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
binder::Status IncrementalService::DataLoaderStub::onStatusChanged(MountId mountId, int newStatus) {
|
|
if (!isValid()) {
|
|
return binder::Status::
|
|
fromServiceSpecificError(-EINVAL, "onStatusChange came to invalid DataLoaderStub");
|
|
}
|
|
if (id() != mountId) {
|
|
LOG(ERROR) << "onStatusChanged: mount ID mismatch: expected " << id()
|
|
<< ", but got: " << mountId;
|
|
return binder::Status::fromServiceSpecificError(-EPERM, "Mount ID mismatch.");
|
|
}
|
|
if (newStatus == IDataLoaderStatusListener::DATA_LOADER_UNAVAILABLE ||
|
|
newStatus == IDataLoaderStatusListener::DATA_LOADER_UNRECOVERABLE) {
|
|
// User-provided status, let's postpone the handling to avoid possible deadlocks.
|
|
mService.addTimedJob(*mService.mTimedQueue, id(), Constants::userStatusDelay,
|
|
[this, newStatus]() { setCurrentStatus(newStatus); });
|
|
return binder::Status::ok();
|
|
}
|
|
|
|
setCurrentStatus(newStatus);
|
|
return binder::Status::ok();
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::setCurrentStatus(int newStatus) {
|
|
compareAndSetCurrentStatus(Constants::anyStatus, newStatus);
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::compareAndSetCurrentStatus(int expectedStatus,
|
|
int newStatus) {
|
|
int oldStatus, oldTargetStatus, newTargetStatus;
|
|
DataLoaderStatusListener listener;
|
|
{
|
|
std::unique_lock lock(mMutex);
|
|
if (mCurrentStatus == newStatus) {
|
|
return;
|
|
}
|
|
if (expectedStatus != Constants::anyStatus && expectedStatus != mCurrentStatus) {
|
|
return;
|
|
}
|
|
|
|
oldStatus = mCurrentStatus;
|
|
oldTargetStatus = mTargetStatus;
|
|
listener = mStatusListener;
|
|
|
|
// Change the status.
|
|
mCurrentStatus = newStatus;
|
|
mCurrentStatusTs = mService.mClock->now();
|
|
|
|
switch (mCurrentStatus) {
|
|
case IDataLoaderStatusListener::DATA_LOADER_UNAVAILABLE:
|
|
// Unavailable, retry.
|
|
setTargetStatusLocked(IDataLoaderStatusListener::DATA_LOADER_STARTED);
|
|
break;
|
|
case IDataLoaderStatusListener::DATA_LOADER_UNRECOVERABLE:
|
|
// Unrecoverable, just unbind.
|
|
setTargetStatusLocked(IDataLoaderStatusListener::DATA_LOADER_DESTROYED);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
newTargetStatus = mTargetStatus;
|
|
}
|
|
|
|
LOG(DEBUG) << "Current status update for DataLoader " << id() << ": " << oldStatus << " -> "
|
|
<< newStatus << " (target " << oldTargetStatus << " -> " << newTargetStatus << ")";
|
|
|
|
if (listener) {
|
|
listener->onStatusChanged(id(), newStatus);
|
|
}
|
|
|
|
fsmStep();
|
|
|
|
mStatusCondition.notify_all();
|
|
}
|
|
|
|
bool IncrementalService::DataLoaderStub::isHealthParamsValid() const {
|
|
return mHealthCheckParams.blockedTimeoutMs > 0 &&
|
|
mHealthCheckParams.blockedTimeoutMs < mHealthCheckParams.unhealthyTimeoutMs;
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::onHealthStatus(const StorageHealthListener& healthListener,
|
|
int healthStatus) {
|
|
LOG(DEBUG) << id() << ": healthStatus: " << healthStatus;
|
|
if (healthListener) {
|
|
healthListener->onHealthStatus(id(), healthStatus);
|
|
}
|
|
mHealthStatus = healthStatus;
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::updateHealthStatus(bool baseline) {
|
|
LOG(DEBUG) << id() << ": updateHealthStatus" << (baseline ? " (baseline)" : "");
|
|
|
|
int healthStatusToReport = -1;
|
|
StorageHealthListener healthListener;
|
|
|
|
{
|
|
std::unique_lock lock(mMutex);
|
|
unregisterFromPendingReads();
|
|
|
|
healthListener = mHealthListener;
|
|
|
|
// Healthcheck depends on timestamp of the oldest pending read.
|
|
// To get it, we need to re-open a pendingReads FD to get a full list of reads.
|
|
// Additionally we need to re-register for epoll with fresh FDs in case there are no
|
|
// reads.
|
|
const auto now = Clock::now();
|
|
const auto kernelTsUs = getOldestPendingReadTs();
|
|
if (baseline) {
|
|
// Updating baseline only on looper/epoll callback, i.e. on new set of pending
|
|
// reads.
|
|
mHealthBase = {now, kernelTsUs};
|
|
}
|
|
|
|
if (kernelTsUs == kMaxBootClockTsUs || mHealthBase.kernelTsUs == kMaxBootClockTsUs ||
|
|
mHealthBase.userTs > now) {
|
|
LOG(DEBUG) << id() << ": No pending reads or invalid base, report Ok and wait.";
|
|
registerForPendingReads();
|
|
healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_OK;
|
|
lock.unlock();
|
|
onHealthStatus(healthListener, healthStatusToReport);
|
|
return;
|
|
}
|
|
|
|
resetHealthControl();
|
|
|
|
// Always make sure the data loader is started.
|
|
setTargetStatusLocked(IDataLoaderStatusListener::DATA_LOADER_STARTED);
|
|
|
|
// Skip any further processing if health check params are invalid.
|
|
if (!isHealthParamsValid()) {
|
|
LOG(DEBUG) << id()
|
|
<< ": Skip any further processing if health check params are invalid.";
|
|
healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_READS_PENDING;
|
|
lock.unlock();
|
|
onHealthStatus(healthListener, healthStatusToReport);
|
|
// Triggering data loader start. This is a one-time action.
|
|
fsmStep();
|
|
return;
|
|
}
|
|
|
|
// Don't schedule timer job less than 500ms in advance.
|
|
static constexpr auto kTolerance = 500ms;
|
|
|
|
const auto blockedTimeout = std::chrono::milliseconds(mHealthCheckParams.blockedTimeoutMs);
|
|
const auto unhealthyTimeout =
|
|
std::chrono::milliseconds(mHealthCheckParams.unhealthyTimeoutMs);
|
|
const auto unhealthyMonitoring =
|
|
std::max(1000ms,
|
|
std::chrono::milliseconds(mHealthCheckParams.unhealthyMonitoringMs));
|
|
|
|
const auto delta = elapsedMsSinceKernelTs(now, kernelTsUs);
|
|
|
|
Milliseconds checkBackAfter;
|
|
if (delta + kTolerance < blockedTimeout) {
|
|
LOG(DEBUG) << id() << ": Report reads pending and wait for blocked status.";
|
|
checkBackAfter = blockedTimeout - delta;
|
|
healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_READS_PENDING;
|
|
} else if (delta + kTolerance < unhealthyTimeout) {
|
|
LOG(DEBUG) << id() << ": Report blocked and wait for unhealthy.";
|
|
checkBackAfter = unhealthyTimeout - delta;
|
|
healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_BLOCKED;
|
|
} else {
|
|
LOG(DEBUG) << id() << ": Report unhealthy and continue monitoring.";
|
|
checkBackAfter = unhealthyMonitoring;
|
|
healthStatusToReport = IStorageHealthListener::HEALTH_STATUS_UNHEALTHY;
|
|
}
|
|
LOG(DEBUG) << id() << ": updateHealthStatus in " << double(checkBackAfter.count()) / 1000.0
|
|
<< "secs";
|
|
mService.addTimedJob(*mService.mTimedQueue, id(), checkBackAfter,
|
|
[this]() { updateHealthStatus(); });
|
|
}
|
|
|
|
// With kTolerance we are expecting these to execute before the next update.
|
|
if (healthStatusToReport != -1) {
|
|
onHealthStatus(healthListener, healthStatusToReport);
|
|
}
|
|
|
|
fsmStep();
|
|
}
|
|
|
|
Milliseconds IncrementalService::DataLoaderStub::elapsedMsSinceKernelTs(TimePoint now,
|
|
BootClockTsUs kernelTsUs) {
|
|
const auto kernelDeltaUs = kernelTsUs - mHealthBase.kernelTsUs;
|
|
const auto userTs = mHealthBase.userTs + std::chrono::microseconds(kernelDeltaUs);
|
|
return std::chrono::duration_cast<Milliseconds>(now - userTs);
|
|
}
|
|
|
|
const incfs::UniqueControl& IncrementalService::DataLoaderStub::initializeHealthControl() {
|
|
if (mHealthPath.empty()) {
|
|
resetHealthControl();
|
|
return mHealthControl;
|
|
}
|
|
if (mHealthControl.pendingReads() < 0) {
|
|
mHealthControl = mService.mIncFs->openMount(mHealthPath);
|
|
}
|
|
if (mHealthControl.pendingReads() < 0) {
|
|
LOG(ERROR) << "Failed to open health control for: " << id() << ", path: " << mHealthPath
|
|
<< "(" << mHealthControl.cmd() << ":" << mHealthControl.pendingReads() << ":"
|
|
<< mHealthControl.logs() << ")";
|
|
}
|
|
return mHealthControl;
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::resetHealthControl() {
|
|
mHealthControl = {};
|
|
}
|
|
|
|
BootClockTsUs IncrementalService::DataLoaderStub::getOldestPendingReadTs() {
|
|
auto result = kMaxBootClockTsUs;
|
|
|
|
const auto& control = initializeHealthControl();
|
|
if (control.pendingReads() < 0) {
|
|
return result;
|
|
}
|
|
|
|
if (mService.mIncFs->waitForPendingReads(control, 0ms, &mLastPendingReads) !=
|
|
android::incfs::WaitResult::HaveData ||
|
|
mLastPendingReads.empty()) {
|
|
// Clear previous pending reads
|
|
mLastPendingReads.clear();
|
|
return result;
|
|
}
|
|
|
|
LOG(DEBUG) << id() << ": pendingReads: fd(" << control.pendingReads() << "), count("
|
|
<< mLastPendingReads.size() << "), block: " << mLastPendingReads.front().block
|
|
<< ", time: " << mLastPendingReads.front().bootClockTsUs
|
|
<< ", uid: " << mLastPendingReads.front().uid;
|
|
|
|
return getOldestTsFromLastPendingReads();
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::registerForPendingReads() {
|
|
const auto pendingReadsFd = mHealthControl.pendingReads();
|
|
if (pendingReadsFd < 0) {
|
|
return;
|
|
}
|
|
|
|
LOG(DEBUG) << id() << ": addFd(pendingReadsFd): " << pendingReadsFd;
|
|
|
|
mService.mLooper->addFd(
|
|
pendingReadsFd, android::Looper::POLL_CALLBACK, android::Looper::EVENT_INPUT,
|
|
[](int, int, void* data) -> int {
|
|
auto self = (DataLoaderStub*)data;
|
|
self->updateHealthStatus(/*baseline=*/true);
|
|
return 0;
|
|
},
|
|
this);
|
|
mService.mLooper->wake();
|
|
}
|
|
|
|
BootClockTsUs IncrementalService::DataLoaderStub::getOldestTsFromLastPendingReads() {
|
|
auto result = kMaxBootClockTsUs;
|
|
for (auto&& pendingRead : mLastPendingReads) {
|
|
result = std::min(result, pendingRead.bootClockTsUs);
|
|
}
|
|
return result;
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::getMetrics(android::os::PersistableBundle* result) {
|
|
const auto duration = elapsedMsSinceOldestPendingRead();
|
|
if (duration >= 0) {
|
|
const auto& kMetricsMillisSinceOldestPendingRead =
|
|
os::incremental::BnIncrementalService::METRICS_MILLIS_SINCE_OLDEST_PENDING_READ();
|
|
result->putLong(String16(kMetricsMillisSinceOldestPendingRead.c_str()), duration);
|
|
}
|
|
const auto& kMetricsStorageHealthStatusCode =
|
|
os::incremental::BnIncrementalService::METRICS_STORAGE_HEALTH_STATUS_CODE();
|
|
result->putInt(String16(kMetricsStorageHealthStatusCode.c_str()), mHealthStatus);
|
|
const auto& kMetricsDataLoaderStatusCode =
|
|
os::incremental::BnIncrementalService::METRICS_DATA_LOADER_STATUS_CODE();
|
|
result->putInt(String16(kMetricsDataLoaderStatusCode.c_str()), mCurrentStatus);
|
|
const auto& kMetricsMillisSinceLastDataLoaderBind =
|
|
os::incremental::BnIncrementalService::METRICS_MILLIS_SINCE_LAST_DATA_LOADER_BIND();
|
|
result->putLong(String16(kMetricsMillisSinceLastDataLoaderBind.c_str()),
|
|
elapsedMcs(mPreviousBindTs, mService.mClock->now()) / 1000);
|
|
const auto& kMetricsDataLoaderBindDelayMillis =
|
|
os::incremental::BnIncrementalService::METRICS_DATA_LOADER_BIND_DELAY_MILLIS();
|
|
result->putLong(String16(kMetricsDataLoaderBindDelayMillis.c_str()),
|
|
mPreviousBindDelay.count());
|
|
}
|
|
|
|
long IncrementalService::DataLoaderStub::elapsedMsSinceOldestPendingRead() {
|
|
const auto oldestPendingReadKernelTs = getOldestTsFromLastPendingReads();
|
|
if (oldestPendingReadKernelTs == kMaxBootClockTsUs) {
|
|
return 0;
|
|
}
|
|
return elapsedMsSinceKernelTs(Clock::now(), oldestPendingReadKernelTs).count();
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::unregisterFromPendingReads() {
|
|
const auto pendingReadsFd = mHealthControl.pendingReads();
|
|
if (pendingReadsFd < 0) {
|
|
return;
|
|
}
|
|
|
|
LOG(DEBUG) << id() << ": removeFd(pendingReadsFd): " << pendingReadsFd;
|
|
|
|
mService.mLooper->removeFd(pendingReadsFd);
|
|
mService.mLooper->wake();
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::setHealthListener(
|
|
const StorageHealthCheckParams& healthCheckParams, StorageHealthListener&& healthListener) {
|
|
std::lock_guard lock(mMutex);
|
|
mHealthCheckParams = healthCheckParams;
|
|
mHealthListener = std::move(healthListener);
|
|
if (!mHealthListener) {
|
|
mHealthCheckParams.blockedTimeoutMs = -1;
|
|
}
|
|
}
|
|
|
|
static std::string toHexString(const RawMetadata& metadata) {
|
|
int n = metadata.size();
|
|
std::string res(n * 2, '\0');
|
|
// Same as incfs::toString(fileId)
|
|
static constexpr char kHexChar[] = "0123456789abcdef";
|
|
for (int i = 0; i < n; ++i) {
|
|
res[i * 2] = kHexChar[(metadata[i] & 0xf0) >> 4];
|
|
res[i * 2 + 1] = kHexChar[(metadata[i] & 0x0f)];
|
|
}
|
|
return res;
|
|
}
|
|
|
|
void IncrementalService::DataLoaderStub::onDump(int fd) {
|
|
dprintf(fd, " dataLoader: {\n");
|
|
dprintf(fd, " currentStatus: %d\n", mCurrentStatus);
|
|
dprintf(fd, " currentStatusTs: %lldmcs\n",
|
|
(long long)(elapsedMcs(mCurrentStatusTs, Clock::now())));
|
|
dprintf(fd, " targetStatus: %d\n", mTargetStatus);
|
|
dprintf(fd, " targetStatusTs: %lldmcs\n",
|
|
(long long)(elapsedMcs(mTargetStatusTs, Clock::now())));
|
|
dprintf(fd, " health: {\n");
|
|
dprintf(fd, " path: %s\n", mHealthPath.c_str());
|
|
dprintf(fd, " base: %lldmcs (%lld)\n",
|
|
(long long)(elapsedMcs(mHealthBase.userTs, Clock::now())),
|
|
(long long)mHealthBase.kernelTsUs);
|
|
dprintf(fd, " blockedTimeoutMs: %d\n", int(mHealthCheckParams.blockedTimeoutMs));
|
|
dprintf(fd, " unhealthyTimeoutMs: %d\n", int(mHealthCheckParams.unhealthyTimeoutMs));
|
|
dprintf(fd, " unhealthyMonitoringMs: %d\n",
|
|
int(mHealthCheckParams.unhealthyMonitoringMs));
|
|
dprintf(fd, " lastPendingReads: \n");
|
|
const auto control = mService.mIncFs->openMount(mHealthPath);
|
|
for (auto&& pendingRead : mLastPendingReads) {
|
|
dprintf(fd, " fileId: %s\n", IncFsWrapper::toString(pendingRead.id).c_str());
|
|
const auto metadata = mService.mIncFs->getMetadata(control, pendingRead.id);
|
|
dprintf(fd, " metadataHex: %s\n", toHexString(metadata).c_str());
|
|
dprintf(fd, " blockIndex: %d\n", pendingRead.block);
|
|
dprintf(fd, " bootClockTsUs: %lld\n", (long long)pendingRead.bootClockTsUs);
|
|
}
|
|
dprintf(fd, " bind: %llds ago (delay: %llds)\n",
|
|
(long long)(elapsedMcs(mPreviousBindTs, mService.mClock->now()) / 1000000),
|
|
(long long)(mPreviousBindDelay.count() / 1000));
|
|
dprintf(fd, " }\n");
|
|
const auto& params = mParams;
|
|
dprintf(fd, " dataLoaderParams: {\n");
|
|
dprintf(fd, " type: %s\n", toString(params.type).c_str());
|
|
dprintf(fd, " packageName: %s\n", params.packageName.c_str());
|
|
dprintf(fd, " className: %s\n", params.className.c_str());
|
|
dprintf(fd, " arguments: %s\n", params.arguments.c_str());
|
|
dprintf(fd, " }\n");
|
|
dprintf(fd, " }\n");
|
|
}
|
|
|
|
void IncrementalService::AppOpsListener::opChanged(int32_t, const String16&) {
|
|
incrementalService.onAppOpChanged(packageName);
|
|
}
|
|
|
|
binder::Status IncrementalService::IncrementalServiceConnector::setStorageParams(
|
|
bool enableReadLogs, int32_t* _aidl_return) {
|
|
*_aidl_return = incrementalService.setStorageParams(storage, enableReadLogs);
|
|
return binder::Status::ok();
|
|
}
|
|
|
|
FileId IncrementalService::idFromMetadata(std::span<const uint8_t> metadata) {
|
|
return IncFs_FileIdFromMetadata({(const char*)metadata.data(), metadata.size()});
|
|
}
|
|
|
|
} // namespace android::incremental
|