Merge "Add annotations to VTS tests" into main
diff --git a/bootstat/bootstat.cpp b/bootstat/bootstat.cpp
index 2d55e5a..d402bf1 100644
--- a/bootstat/bootstat.cpp
+++ b/bootstat/bootstat.cpp
@@ -830,7 +830,7 @@
return subReason;
}
-bool addKernelPanicSubReason(const pstoreConsole& console, std::string& ret) {
+void addKernelPanicSubReason(const pstoreConsole& console, std::string& ret) {
// Check for kernel panic types to refine information
if ((console.rfind("SysRq : Trigger a crash") != std::string::npos) ||
(console.rfind("PC is at sysrq_handle_crash+") != std::string::npos)) {
@@ -842,63 +842,61 @@
if (pos != std::string::npos) {
ret += "," + getSubreason(console, pos + strlen(sysrqSubreason), /* quoted */ true);
}
- return true;
+ return;
}
if (console.rfind("Unable to handle kernel NULL pointer dereference at virtual address") !=
std::string::npos) {
ret = "kernel_panic,null";
- return true;
+ return;
}
if (console.rfind("Kernel BUG at ") != std::string::npos) {
ret = "kernel_panic,bug";
- return true;
+ return;
}
std::string panic("Kernel panic - not syncing: ");
auto pos = console.rfind(panic);
- if (pos != std::string::npos) {
- static const std::vector<std::pair<const std::string, const std::string>> panicReasons = {
- {"Out of memory", "oom"},
- {"out of memory", "oom"},
- {"Oh boy, that early out of memory", "oom"}, // omg
- {"BUG!", "bug"},
- {"hung_task: blocked tasks", "hung"},
- {"audit: ", "audit"},
- {"scheduling while atomic", "atomic"},
- {"Attempted to kill init!", "init"},
- {"Requested init", "init"},
- {"No working init", "init"},
- {"Could not decompress init", "init"},
- {"RCU Stall", "hung,rcu"},
- {"stack-protector", "stack"},
- {"kernel stack overflow", "stack"},
- {"Corrupt kernel stack", "stack"},
- {"low stack detected", "stack"},
- {"corrupted stack end", "stack"},
- {"subsys-restart: Resetting the SoC - modem crashed.", "modem"},
- {"subsys-restart: Resetting the SoC - adsp crashed.", "adsp"},
- {"subsys-restart: Resetting the SoC - dsps crashed.", "dsps"},
- {"subsys-restart: Resetting the SoC - wcnss crashed.", "wcnss"},
- };
+ if (pos == std::string::npos) return;
- ret = "kernel_panic";
- for (auto& s : panicReasons) {
- if (console.find(panic + s.first, pos) != std::string::npos) {
- ret += "," + s.second;
- return true;
- }
+ static const std::vector<std::pair<const std::string, const std::string>> panicReasons = {
+ {"Out of memory", "oom"},
+ {"out of memory", "oom"},
+ {"Oh boy, that early out of memory", "oom"}, // omg
+ {"BUG!", "bug"},
+ {"hung_task: blocked tasks", "hung"},
+ {"audit: ", "audit"},
+ {"scheduling while atomic", "atomic"},
+ {"Attempted to kill init!", "init"},
+ {"Requested init", "init"},
+ {"No working init", "init"},
+ {"Could not decompress init", "init"},
+ {"RCU Stall", "hung,rcu"},
+ {"stack-protector", "stack"},
+ {"kernel stack overflow", "stack"},
+ {"Corrupt kernel stack", "stack"},
+ {"low stack detected", "stack"},
+ {"corrupted stack end", "stack"},
+ {"subsys-restart: Resetting the SoC - modem crashed.", "modem"},
+ {"subsys-restart: Resetting the SoC - adsp crashed.", "adsp"},
+ {"subsys-restart: Resetting the SoC - dsps crashed.", "dsps"},
+ {"subsys-restart: Resetting the SoC - wcnss crashed.", "wcnss"},
+ };
+
+ ret = "kernel_panic";
+ for (auto& s : panicReasons) {
+ if (console.find(panic + s.first, pos) != std::string::npos) {
+ ret += "," + s.second;
+ return;
}
- auto reason = getSubreason(console, pos + panic.length(), /* newline */ false);
- if (reason.length() > 3) {
- ret += "," + reason;
- }
- return true;
}
- return false;
+ auto reason = getSubreason(console, pos + panic.length(), /* newline */ false);
+ if (reason.length() > 3) {
+ ret += "," + reason;
+ }
}
-bool addKernelPanicSubReason(const std::string& content, std::string& ret) {
- return addKernelPanicSubReason(pstoreConsole(content), ret);
+void addKernelPanicSubReason(const std::string& content, std::string& ret) {
+ addKernelPanicSubReason(pstoreConsole(content), ret);
}
const char system_reboot_reason_property[] = "sys.boot.reason";
@@ -1079,12 +1077,7 @@
}
// Check for kernel panics, allowed to override reboot command.
- if (!addKernelPanicSubReason(console, ret) &&
- // check for long-press power down
- ((console.rfind("Power held for ") != std::string::npos) ||
- (console.rfind("charger: [") != std::string::npos))) {
- ret = "cold";
- }
+ (void)addKernelPanicSubReason(console, ret);
}
// TODO: use the HAL to get battery level (http://b/77725702).
diff --git a/debuggerd/Android.bp b/debuggerd/Android.bp
index 7d20995..439218d 100644
--- a/debuggerd/Android.bp
+++ b/debuggerd/Android.bp
@@ -21,8 +21,11 @@
local_include_dirs: ["include"],
product_variables: {
debuggable: {
- cflags: ["-UANDROID_DEBUGGABLE", "-DANDROID_DEBUGGABLE=1"],
- }
+ cflags: [
+ "-UANDROID_DEBUGGABLE",
+ "-DANDROID_DEBUGGABLE=1",
+ ],
+ },
},
}
@@ -256,7 +259,7 @@
],
static_libs: [
- "libdexfile_support", // libunwindstack dependency
+ "libdexfile_support", // libunwindstack dependency
"libunwindstack",
"liblzma",
"libbase",
@@ -299,7 +302,7 @@
},
android: {
runtime_libs: [
- "libdexfile", // libdexfile_support dependency
+ "libdexfile", // libdexfile_support dependency
],
},
},
@@ -494,7 +497,7 @@
header_libs: [
"bionic_libc_platform_headers",
- "libdebuggerd_common_headers"
+ "libdebuggerd_common_headers",
],
static_libs: [
@@ -552,7 +555,6 @@
},
}
-
// This installs the "other" architecture (so 32-bit on 64-bit device).
prebuilt_etc {
name: "crash_dump.policy_other",
diff --git a/debuggerd/client/debuggerd_client.cpp b/debuggerd/client/debuggerd_client.cpp
index bd1e91d..af1bb81 100644
--- a/debuggerd/client/debuggerd_client.cpp
+++ b/debuggerd/client/debuggerd_client.cpp
@@ -116,7 +116,6 @@
bool debuggerd_trigger_dump(pid_t tid, DebuggerdDumpType dump_type, unsigned int timeout_ms,
unique_fd output_fd) {
- pid_t pid = tid;
if (dump_type == kDebuggerdJavaBacktrace) {
// Java dumps always get sent to the tgid, so we need to resolve our tid to a tgid.
android::procinfo::ProcessInfo procinfo;
@@ -125,10 +124,10 @@
log_error(output_fd, 0, "failed to get process info: %s", error.c_str());
return false;
}
- pid = procinfo.pid;
+ tid = procinfo.pid;
}
- LOG(INFO) << TAG "started dumping process " << pid;
+ LOG(INFO) << TAG "started dumping process " << tid;
// Rather than try to deal with poll() all the way through the flow, we update
// the socket timeout between each step (and only use poll() during the final
@@ -172,7 +171,7 @@
InterceptRequest req = {
.dump_type = dump_type,
- .pid = pid,
+ .pid = tid,
};
// Create an intermediate pipe to pass to the other end.
@@ -235,8 +234,8 @@
// Send the signal.
const int signal = (dump_type == kDebuggerdJavaBacktrace) ? SIGQUIT : BIONIC_SIGNAL_DEBUGGER;
sigval val = {.sival_int = (dump_type == kDebuggerdNativeBacktrace) ? 1 : 0};
- if (sigqueue(pid, signal, val) != 0) {
- log_error(output_fd, errno, "failed to send signal to pid %d", pid);
+ if (sigqueue(tid, signal, val) != 0) {
+ log_error(output_fd, errno, "failed to send signal to pid %d", tid);
return false;
}
@@ -299,7 +298,7 @@
}
}
- LOG(INFO) << TAG "done dumping process " << pid;
+ LOG(INFO) << TAG "done dumping process " << tid;
return true;
}
diff --git a/debuggerd/crash_dump.cpp b/debuggerd/crash_dump.cpp
index 0899111..a23a269 100644
--- a/debuggerd/crash_dump.cpp
+++ b/debuggerd/crash_dump.cpp
@@ -322,6 +322,7 @@
process_info->scudo_ring_buffer = crash_info->data.d.scudo_ring_buffer;
process_info->scudo_ring_buffer_size = crash_info->data.d.scudo_ring_buffer_size;
*recoverable_gwp_asan_crash = crash_info->data.d.recoverable_gwp_asan_crash;
+ process_info->crash_detail_page = crash_info->data.d.crash_detail_page;
FALLTHROUGH_INTENDED;
case 1:
case 2:
diff --git a/debuggerd/crasher/Android.bp b/debuggerd/crasher/Android.bp
index fe1689c..4c6a400 100644
--- a/debuggerd/crasher/Android.bp
+++ b/debuggerd/crasher/Android.bp
@@ -15,7 +15,7 @@
"-fstack-protector-all",
"-Wno-date-time",
],
- tidy: false, // crasher.cpp tests many memory access errors
+ tidy: false, // crasher.cpp tests many memory access errors
srcs: ["crasher.cpp"],
arch: {
arm: {
diff --git a/debuggerd/debuggerd_test.cpp b/debuggerd/debuggerd_test.cpp
index c0522aa..5cc2b3a 100644
--- a/debuggerd/debuggerd_test.cpp
+++ b/debuggerd/debuggerd_test.cpp
@@ -37,6 +37,7 @@
#include <string>
#include <thread>
+#include <android/crash_detail.h>
#include <android/dlext.h>
#include <android/fdsan.h>
#include <android/set_abort_message.h>
@@ -939,6 +940,233 @@
ASSERT_MATCH(result, R"(Abort message: 'x{4045}')");
}
+static char g_crash_detail_value_changes[] = "crash_detail_value";
+static char g_crash_detail_value[] = "crash_detail_value";
+static char g_crash_detail_value2[] = "crash_detail_value2";
+
+inline crash_detail_t* _Nullable android_register_crash_detail_strs(const char* _Nonnull name,
+ const char* _Nonnull data) {
+ return android_crash_detail_register(name, strlen(name), data, strlen(data));
+}
+
+TEST_F(CrasherTest, crash_detail_single) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ android_register_crash_detail_strs("CRASH_DETAIL_NAME", g_crash_detail_value);
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME: 'crash_detail_value')");
+}
+
+TEST_F(CrasherTest, crash_detail_replace_data) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ auto *cd = android_register_crash_detail_strs("CRASH_DETAIL_NAME", "original_data");
+ android_crash_detail_replace_data(cd, "new_data", strlen("new_data"));
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME: 'new_data')");
+ // Ensure the old one no longer shows up, i.e. that we actually replaced
+ // it, not added a new one.
+ ASSERT_NOT_MATCH(result, R"(CRASH_DETAIL_NAME: 'original_data')");
+}
+
+TEST_F(CrasherTest, crash_detail_replace_name) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ auto *cd = android_register_crash_detail_strs("old_name", g_crash_detail_value);
+ android_crash_detail_replace_name(cd, "new_name", strlen("new_name"));
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_MATCH(result, R"(new_name: 'crash_detail_value')");
+ // Ensure the old one no longer shows up, i.e. that we actually replaced
+ // it, not added a new one.
+ ASSERT_NOT_MATCH(result, R"(old_name: 'crash_detail_value')");
+}
+
+TEST_F(CrasherTest, crash_detail_single_byte_name) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ android_register_crash_detail_strs("CRASH_DETAIL_NAME\1", g_crash_detail_value);
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME\\1: 'crash_detail_value')");
+}
+
+
+TEST_F(CrasherTest, crash_detail_single_bytes) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ android_crash_detail_register("CRASH_DETAIL_NAME", strlen("CRASH_DETAIL_NAME"), "\1",
+ sizeof("\1"));
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME: '\\1\\0')");
+}
+
+TEST_F(CrasherTest, crash_detail_mixed) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ const char data[] = "helloworld\1\255\3";
+ android_register_crash_detail_strs("CRASH_DETAIL_NAME", data);
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME: 'helloworld\\1\\255\\3')");
+}
+
+TEST_F(CrasherTest, crash_detail_many) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ for (int i = 0; i < 1000; ++i) {
+ std::string name = "CRASH_DETAIL_NAME" + std::to_string(i);
+ std::string value = "CRASH_DETAIL_VALUE" + std::to_string(i);
+ auto* h = android_register_crash_detail_strs(name.data(), value.data());
+ android_crash_detail_unregister(h);
+ }
+
+ android_register_crash_detail_strs("FINAL_NAME", "FINAL_VALUE");
+ android_register_crash_detail_strs("FINAL_NAME2", "FINAL_VALUE2");
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_NOT_MATCH(result, "CRASH_DETAIL_NAME");
+ ASSERT_NOT_MATCH(result, "CRASH_DETAIL_VALUE");
+ ASSERT_MATCH(result, R"(FINAL_NAME: 'FINAL_VALUE')");
+ ASSERT_MATCH(result, R"(FINAL_NAME2: 'FINAL_VALUE2')");
+}
+
+TEST_F(CrasherTest, crash_detail_single_changes) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ android_register_crash_detail_strs("CRASH_DETAIL_NAME", g_crash_detail_value_changes);
+ g_crash_detail_value_changes[0] = 'C';
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME: 'Crash_detail_value')");
+}
+
+TEST_F(CrasherTest, crash_detail_multiple) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ android_register_crash_detail_strs("CRASH_DETAIL_NAME", g_crash_detail_value);
+ android_register_crash_detail_strs("CRASH_DETAIL_NAME2", g_crash_detail_value2);
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME: 'crash_detail_value')");
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME2: 'crash_detail_value2')");
+}
+
+TEST_F(CrasherTest, crash_detail_remove) {
+ int intercept_result;
+ unique_fd output_fd;
+ StartProcess([]() {
+ auto* detail1 = android_register_crash_detail_strs("CRASH_DETAIL_NAME", g_crash_detail_value);
+ android_crash_detail_unregister(detail1);
+ android_register_crash_detail_strs("CRASH_DETAIL_NAME2", g_crash_detail_value2);
+ abort();
+ });
+ StartIntercept(&output_fd);
+ FinishCrasher();
+ AssertDeath(SIGABRT);
+ FinishIntercept(&intercept_result);
+
+ ASSERT_EQ(1, intercept_result) << "tombstoned reported failure";
+
+ std::string result;
+ ConsumeFd(std::move(output_fd), &result);
+ ASSERT_NOT_MATCH(result, R"(CRASH_DETAIL_NAME: 'crash_detail_value')");
+ ASSERT_MATCH(result, R"(CRASH_DETAIL_NAME2: 'crash_detail_value2')");
+}
+
TEST_F(CrasherTest, abort_message_newline_trimmed) {
int intercept_result;
unique_fd output_fd;
@@ -2419,7 +2647,7 @@
match_str += format_full_pointer(crash_uptr);
ASSERT_MATCH(result, match_str);
- ASSERT_MATCH(result, R"(\nmemory map \(.*\): \(fault address prefixed with --->)\n)");
+ ASSERT_MATCH(result, R"(\nmemory map \(.*\): \(fault address prefixed with --->\)\n)");
// Verifies that the fault address error message is at the end of the
// maps section. To do this, the check below looks for the start of the
@@ -2471,7 +2699,7 @@
match_str += format_full_pointer(reinterpret_cast<uintptr_t>(middle_ptr));
ASSERT_MATCH(result, match_str);
- ASSERT_MATCH(result, R"(\nmemory map \(.*\): \(fault address prefixed with --->)\n)");
+ ASSERT_MATCH(result, R"(\nmemory map \(.*\): \(fault address prefixed with --->\)\n)");
match_str = android::base::StringPrintf(
R"( %s.*\n--->Fault address falls at %s between mapped regions\n %s)",
@@ -2509,7 +2737,7 @@
match_str += format_full_pointer(reinterpret_cast<uintptr_t>(ptr));
ASSERT_MATCH(result, match_str);
- ASSERT_MATCH(result, R"(\nmemory map \(.*\): \(fault address prefixed with --->)\n)");
+ ASSERT_MATCH(result, R"(\nmemory map \(.*\): \(fault address prefixed with --->\)\n)");
match_str = android::base::StringPrintf(R"(\n--->%s.*\n)", format_pointer(ptr).c_str());
ASSERT_MATCH(result, match_str);
diff --git a/debuggerd/handler/debuggerd_handler.cpp b/debuggerd/handler/debuggerd_handler.cpp
index ea07ce2..141723b 100644
--- a/debuggerd/handler/debuggerd_handler.cpp
+++ b/debuggerd/handler/debuggerd_handler.cpp
@@ -397,6 +397,7 @@
ASSERT_SAME_OFFSET(scudo_ring_buffer_size, scudo_ring_buffer_size);
ASSERT_SAME_OFFSET(scudo_stack_depot_size, scudo_stack_depot_size);
ASSERT_SAME_OFFSET(recoverable_gwp_asan_crash, recoverable_gwp_asan_crash);
+ ASSERT_SAME_OFFSET(crash_detail_page, crash_detail_page);
#undef ASSERT_SAME_OFFSET
iovs[3] = {.iov_base = &thread_info->process_info,
diff --git a/debuggerd/include/debuggerd/client.h b/debuggerd/include/debuggerd/client.h
index b7284b0..e7401cc 100644
--- a/debuggerd/include/debuggerd/client.h
+++ b/debuggerd/include/debuggerd/client.h
@@ -26,7 +26,7 @@
// Trigger a dump of specified process to output_fd.
// output_fd is consumed, timeout of 0 will wait forever.
-bool debuggerd_trigger_dump(pid_t pid, enum DebuggerdDumpType dump_type, unsigned int timeout_ms,
+bool debuggerd_trigger_dump(pid_t tid, enum DebuggerdDumpType dump_type, unsigned int timeout_ms,
android::base::unique_fd output_fd);
int dump_backtrace_to_file(pid_t tid, enum DebuggerdDumpType dump_type, int output_fd);
diff --git a/debuggerd/include/debuggerd/handler.h b/debuggerd/include/debuggerd/handler.h
index de12fc6..c18cf6e 100644
--- a/debuggerd/include/debuggerd/handler.h
+++ b/debuggerd/include/debuggerd/handler.h
@@ -33,6 +33,8 @@
struct AllocationMetadata;
}; // namespace gwp_asan
+struct crash_detail_page_t;
+
// When updating this data structure, CrashInfoDataDynamic and the code in
// ReadCrashInfo() must also be updated.
struct __attribute__((packed)) debugger_process_info {
@@ -46,6 +48,7 @@
size_t scudo_ring_buffer_size;
size_t scudo_stack_depot_size;
bool recoverable_gwp_asan_crash;
+ struct crash_detail_page_t* crash_detail_page;
};
// GWP-ASan calbacks to support the recoverable mode. Separate from the
diff --git a/debuggerd/libdebuggerd/include/libdebuggerd/types.h b/debuggerd/libdebuggerd/include/libdebuggerd/types.h
index 075b12c..4d69658 100644
--- a/debuggerd/libdebuggerd/include/libdebuggerd/types.h
+++ b/debuggerd/libdebuggerd/include/libdebuggerd/types.h
@@ -56,4 +56,5 @@
bool has_fault_address = false;
uintptr_t untagged_fault_address = 0;
uintptr_t maybe_tagged_fault_address = 0;
+ uintptr_t crash_detail_page = 0;
};
diff --git a/debuggerd/libdebuggerd/test/tombstone_proto_to_text_test.cpp b/debuggerd/libdebuggerd/test/tombstone_proto_to_text_test.cpp
index ac92ac0..a4c08a4 100644
--- a/debuggerd/libdebuggerd/test/tombstone_proto_to_text_test.cpp
+++ b/debuggerd/libdebuggerd/test/tombstone_proto_to_text_test.cpp
@@ -118,3 +118,19 @@
"LOG pac_enabled_keys: 0000000000001009 \\(PR_PAC_APIAKEY, PR_PAC_APDBKEY, unknown "
"0x1000\\)\\n");
}
+
+TEST_F(TombstoneProtoToTextTest, crash_detail_string) {
+ auto* crash_detail = tombstone_->add_crash_details();
+ crash_detail->set_name("CRASH_DETAIL_NAME");
+ crash_detail->set_data("crash_detail_value");
+ ProtoToString();
+ EXPECT_MATCH(text_, "(CRASH_DETAIL_NAME: 'crash_detail_value')");
+}
+
+TEST_F(TombstoneProtoToTextTest, crash_detail_bytes) {
+ auto* crash_detail = tombstone_->add_crash_details();
+ crash_detail->set_name("CRASH_DETAIL_NAME");
+ crash_detail->set_data("helloworld\1\255\3");
+ ProtoToString();
+ EXPECT_MATCH(text_, R"(CRASH_DETAIL_NAME: 'helloworld\\1\\255\\3')");
+}
diff --git a/debuggerd/libdebuggerd/tombstone_proto.cpp b/debuggerd/libdebuggerd/tombstone_proto.cpp
index 744bfab..74f9a8c 100644
--- a/debuggerd/libdebuggerd/tombstone_proto.cpp
+++ b/debuggerd/libdebuggerd/tombstone_proto.cpp
@@ -48,8 +48,10 @@
#include <android-base/unique_fd.h>
#include <android/log.h>
+#include <android/set_abort_message.h>
#include <bionic/macros.h>
#include <bionic/reserved_signals.h>
+#include <bionic/crash_detail_internal.h>
#include <log/log.h>
#include <log/log_read.h>
#include <log/logprint.h>
@@ -94,6 +96,11 @@
static std::optional<std::string> get_stack_overflow_cause(uint64_t fault_addr, uint64_t sp,
unwindstack::Maps* maps) {
+ // Under stack MTE the stack pointer and/or the fault address can be tagged.
+ // In order to calculate deltas between them, strip off the tags off both
+ // addresses.
+ fault_addr = untag_address(fault_addr);
+ sp = untag_address(sp);
static constexpr uint64_t kMaxDifferenceBytes = 256;
uint64_t difference;
if (sp >= fault_addr) {
@@ -251,6 +258,46 @@
}
}
+static void dump_crash_details(Tombstone* tombstone,
+ std::shared_ptr<unwindstack::Memory>& process_memory,
+ const ProcessInfo& process_info) {
+ uintptr_t address = process_info.crash_detail_page;
+ while (address) {
+ struct crash_detail_page_t page;
+ if (!process_memory->ReadFully(address, &page, sizeof(page))) {
+ async_safe_format_log(ANDROID_LOG_ERROR, LOG_TAG, "failed to read crash detail page: %m");
+ break;
+ }
+ if (page.used > kNumCrashDetails) {
+ async_safe_format_log(ANDROID_LOG_ERROR, LOG_TAG, "crash detail: page corrupted");
+ break;
+ }
+ for (size_t i = 0; i < page.used; ++i) {
+ const crash_detail_t& crash_detail = page.crash_details[i];
+ if (!crash_detail.data) {
+ continue;
+ }
+ std::string name(crash_detail.name_size, '\0');
+ if (!process_memory->ReadFully(reinterpret_cast<uintptr_t>(crash_detail.name), name.data(),
+ crash_detail.name_size)) {
+ async_safe_format_log(ANDROID_LOG_ERROR, LOG_TAG, "crash detail: failed to read name: %m");
+ continue;
+ }
+ std::string data(crash_detail.data_size, '\0');
+ if (!process_memory->ReadFully(reinterpret_cast<uintptr_t>(crash_detail.data), data.data(),
+ crash_detail.data_size)) {
+ async_safe_format_log(ANDROID_LOG_ERROR, LOG_TAG,
+ "crash detail: failed to read data for %s: %m", name.c_str());
+ continue;
+ }
+ auto* proto_detail = tombstone->add_crash_details();
+ proto_detail->set_name(name);
+ proto_detail->set_data(data);
+ }
+ address = reinterpret_cast<uintptr_t>(page.prev);
+ }
+}
+
static void dump_abort_message(Tombstone* tombstone,
std::shared_ptr<unwindstack::Memory>& process_memory,
const ProcessInfo& process_info) {
@@ -698,7 +745,7 @@
*result.mutable_signal_info() = sig;
dump_abort_message(&result, unwinder->GetProcessMemory(), process_info);
-
+ dump_crash_details(&result, unwinder->GetProcessMemory(), process_info);
// Dump the main thread, but save the memory around the registers.
dump_thread(&result, unwinder, main_thread, /* memory_dump */ true);
diff --git a/debuggerd/libdebuggerd/tombstone_proto_to_text.cpp b/debuggerd/libdebuggerd/tombstone_proto_to_text.cpp
index ad91320..cefa2d6 100644
--- a/debuggerd/libdebuggerd/tombstone_proto_to_text.cpp
+++ b/debuggerd/libdebuggerd/tombstone_proto_to_text.cpp
@@ -18,7 +18,9 @@
#include <inttypes.h>
+#include <charconv>
#include <functional>
+#include <limits>
#include <set>
#include <string>
#include <unordered_set>
@@ -425,6 +427,27 @@
}
}
+static std::string oct_encode(const std::string& data) {
+ std::string oct_encoded;
+ oct_encoded.reserve(data.size());
+
+ // N.B. the unsigned here is very important, otherwise e.g. \255 would render as
+ // \-123 (and overflow our buffer).
+ for (unsigned char c : data) {
+ if (isprint(c)) {
+ oct_encoded += c;
+ } else {
+ std::string oct_digits("\\\0\0\0", 4);
+ // char is encodable in 3 oct digits
+ static_assert(std::numeric_limits<unsigned char>::max() <= 8 * 8 * 8);
+ auto [ptr, ec] = std::to_chars(oct_digits.data() + 1, oct_digits.data() + 4, c, 8);
+ oct_digits.resize(ptr - oct_digits.data());
+ oct_encoded += oct_digits;
+ }
+ }
+ return oct_encoded;
+}
+
static void print_main_thread(CallbackType callback, const Tombstone& tombstone,
const Thread& thread) {
print_thread_header(callback, tombstone, thread, true);
@@ -468,6 +491,12 @@
CBL("Abort message: '%s'", tombstone.abort_message().c_str());
}
+ for (const auto& crash_detail : tombstone.crash_details()) {
+ std::string oct_encoded_name = oct_encode(crash_detail.name());
+ std::string oct_encoded_data = oct_encode(crash_detail.data());
+ CBL("Extra crash detail: %s: '%s'", oct_encoded_name.c_str(), oct_encoded_data.c_str());
+ }
+
print_thread_registers(callback, tombstone, thread, true);
if (is_async_mte_crash) {
CBL("Note: This crash is a delayed async MTE crash. Memory corruption has occurred");
diff --git a/debuggerd/proto/Android.bp b/debuggerd/proto/Android.bp
index 7be5d61..7b9e780 100644
--- a/debuggerd/proto/Android.bp
+++ b/debuggerd/proto/Android.bp
@@ -52,4 +52,3 @@
sdk_version: "current",
static_libs: ["libprotobuf-java-lite"],
}
-
diff --git a/debuggerd/proto/tombstone.proto b/debuggerd/proto/tombstone.proto
index 49865a2..214cbfb 100644
--- a/debuggerd/proto/tombstone.proto
+++ b/debuggerd/proto/tombstone.proto
@@ -15,6 +15,11 @@
// NOTE TO OEMS:
// If you add custom fields to this proto, do not use numbers in the reserved range.
+message CrashDetail {
+ bytes name = 1;
+ bytes data = 2;
+}
+
message Tombstone {
Architecture arch = 1;
string build_fingerprint = 2;
@@ -33,6 +38,7 @@
Signal signal_info = 10;
string abort_message = 14;
+ repeated CrashDetail crash_details = 21;
repeated Cause causes = 15;
map<uint32, Thread> threads = 16;
@@ -40,7 +46,7 @@
repeated LogBuffer log_buffers = 18;
repeated FD open_fds = 19;
- reserved 21 to 999;
+ reserved 22 to 999;
}
enum Architecture {
diff --git a/debuggerd/protocol.h b/debuggerd/protocol.h
index 793428a..d3fc15e 100644
--- a/debuggerd/protocol.h
+++ b/debuggerd/protocol.h
@@ -101,6 +101,7 @@
size_t scudo_ring_buffer_size;
size_t scudo_stack_depot_size;
bool recoverable_gwp_asan_crash;
+ uintptr_t crash_detail_page;
};
struct __attribute__((__packed__)) CrashInfo {
diff --git a/debuggerd/rust/tombstoned_client/Android.bp b/debuggerd/rust/tombstoned_client/Android.bp
index 2007f39..bf19bb7 100644
--- a/debuggerd/rust/tombstoned_client/Android.bp
+++ b/debuggerd/rust/tombstoned_client/Android.bp
@@ -8,7 +8,7 @@
"wrapper.cpp",
],
generated_sources: [
- "libtombstoned_client_rust_bridge_code"
+ "libtombstoned_client_rust_bridge_code",
],
header_libs: [
"libbase_headers",
diff --git a/debuggerd/seccomp_policy/crash_dump.arm64.policy b/debuggerd/seccomp_policy/crash_dump.arm64.policy
index adf8738..c5d10d6 100644
--- a/debuggerd/seccomp_policy/crash_dump.arm64.policy
+++ b/debuggerd/seccomp_policy/crash_dump.arm64.policy
@@ -28,11 +28,11 @@
rt_tgsigqueueinfo: 1
prctl: arg0 == PR_GET_NO_NEW_PRIVS || arg0 == 0x53564d41 || arg0 == PR_PAC_RESET_KEYS || arg0 == 56 || arg0 == 61
madvise: 1
-mprotect: arg2 in 0x1|0x2
+mprotect: arg2 in 0x1|0x2|0x20
munmap: 1
getuid: 1
fstat: 1
-mmap: arg2 in 0x1|0x2
+mmap: arg2 in 0x1|0x2|0x20
geteuid: 1
getgid: 1
getegid: 1
diff --git a/debuggerd/seccomp_policy/crash_dump.policy.def b/debuggerd/seccomp_policy/crash_dump.policy.def
index 972a575..dc751da 100644
--- a/debuggerd/seccomp_policy/crash_dump.policy.def
+++ b/debuggerd/seccomp_policy/crash_dump.policy.def
@@ -25,8 +25,8 @@
faccessat: 1
recvmsg: 1
recvfrom: 1
-sysinfo: 1
setsockopt: 1
+sysinfo: 1
process_vm_readv: 1
@@ -53,20 +53,29 @@
#if 0
libminijail on vendor partitions older than P does not have constants from <sys/mman.h>.
-Define the values of PROT_READ and PROT_WRITE ourselves to maintain backwards compatibility.
+Define values for PROT_READ, PROT_WRITE and PROT_MTE ourselves to maintain backwards compatibility.
#else
#define PROT_READ 0x1
#define PROT_WRITE 0x2
+#define PROT_MTE 0x20
#endif
madvise: 1
+#if defined(__aarch64__)
+mprotect: arg2 in PROT_READ|PROT_WRITE|PROT_MTE
+#else
mprotect: arg2 in PROT_READ|PROT_WRITE
+#endif
munmap: 1
#if defined(__LP64__)
getuid: 1
fstat: 1
+#if defined(__aarch64__)
+mmap: arg2 in PROT_READ|PROT_WRITE|PROT_MTE
+#else
mmap: arg2 in PROT_READ|PROT_WRITE
+#endif
#else
getuid32: 1
fstat64: 1
diff --git a/debuggerd/test_permissive_mte/Android.bp b/debuggerd/test_permissive_mte/Android.bp
index d3f7520..0ad3243 100644
--- a/debuggerd/test_permissive_mte/Android.bp
+++ b/debuggerd/test_permissive_mte/Android.bp
@@ -17,22 +17,28 @@
}
cc_binary {
- name: "mte_crash",
- tidy: false,
- srcs: ["mte_crash.cpp"],
- sanitize: {
- memtag_heap: true,
- diag: {
- memtag_heap: true,
+ name: "mte_crash",
+ tidy: false,
+ srcs: ["mte_crash.cpp"],
+ sanitize: {
+ memtag_heap: true,
+ diag: {
+ memtag_heap: true,
+ },
},
- },
}
java_test_host {
name: "permissive_mte_test",
libs: ["tradefed"],
- static_libs: ["frameworks-base-hostutils", "cts-install-lib-host"],
- srcs: ["src/**/PermissiveMteTest.java", ":libtombstone_proto-src"],
+ static_libs: [
+ "frameworks-base-hostutils",
+ "cts-install-lib-host",
+ ],
+ srcs: [
+ "src/**/PermissiveMteTest.java",
+ ":libtombstone_proto-src",
+ ],
data: [":mte_crash"],
test_config: "AndroidTest.xml",
test_suites: ["general-tests"],
diff --git a/debuggerd/tombstoned/tombstoned.cpp b/debuggerd/tombstoned/tombstoned.cpp
index cf7904f..e2a67a2 100644
--- a/debuggerd/tombstoned/tombstoned.cpp
+++ b/debuggerd/tombstoned/tombstoned.cpp
@@ -61,7 +61,6 @@
struct CrashArtifact {
unique_fd fd;
- std::optional<std::string> temporary_path;
static CrashArtifact devnull() {
CrashArtifact result;
@@ -145,16 +144,13 @@
std::optional<std::string> path;
result.fd.reset(openat(dir_fd_, ".", O_WRONLY | O_APPEND | O_TMPFILE | O_CLOEXEC, 0660));
if (result.fd == -1) {
- // We might not have O_TMPFILE. Try creating with an arbitrary filename instead.
- static size_t counter = 0;
- std::string tmp_filename = StringPrintf(".temporary%zu", counter++);
- result.fd.reset(openat(dir_fd_, tmp_filename.c_str(),
- O_WRONLY | O_APPEND | O_CREAT | O_TRUNC | O_CLOEXEC, 0660));
- if (result.fd == -1) {
- PLOG(FATAL) << "failed to create temporary tombstone in " << dir_path_;
- }
+ PLOG(FATAL) << "failed to create temporary tombstone in " << dir_path_;
+ }
- result.temporary_path = std::move(tmp_filename);
+ // We need to fchmodat after creating to avoid getting the umask applied.
+ std::string fd_path = StringPrintf("/proc/self/fd/%d", result.fd.get());
+ if (fchmodat(dir_fd_, fd_path.c_str(), 0664, 0) != 0) {
+ PLOG(ERROR) << "Failed to make tombstone world-readable";
}
return std::move(result);
@@ -417,6 +413,7 @@
return false;
}
+ // This fd is created inside of dirfd in CrashQueue::create_temporary_file.
std::string fd_path = StringPrintf("/proc/self/fd/%d", fd.get());
rc = linkat(AT_FDCWD, fd_path.c_str(), dirfd.get(), path.c_str(), AT_SYMLINK_FOLLOW);
if (rc != 0) {
@@ -471,20 +468,6 @@
rename_tombstone_fd(crash->output.proto->fd, queue->dir_fd(), *paths.proto);
}
}
-
- // If we don't have O_TMPFILE, we need to clean up after ourselves.
- if (crash->output.text.temporary_path) {
- rc = unlinkat(queue->dir_fd().get(), crash->output.text.temporary_path->c_str(), 0);
- if (rc != 0) {
- PLOG(ERROR) << "failed to unlink temporary tombstone at " << paths.text;
- }
- }
- if (crash->output.proto && crash->output.proto->temporary_path) {
- rc = unlinkat(queue->dir_fd().get(), crash->output.proto->temporary_path->c_str(), 0);
- if (rc != 0) {
- PLOG(ERROR) << "failed to unlink temporary proto tombstone";
- }
- }
}
static void crash_completed_cb(evutil_socket_t sockfd, short ev, void* arg) {
diff --git a/fastboot/usb_linux.cpp b/fastboot/usb_linux.cpp
index 37bb304..03af8f7 100644
--- a/fastboot/usb_linux.cpp
+++ b/fastboot/usb_linux.cpp
@@ -83,7 +83,18 @@
// be reliable.
// 256KiB seems to work, but 1MiB bulk transfers lock up my z620 with a 3.13
// kernel.
-#define MAX_USBFS_BULK_SIZE (16 * 1024)
+// 128KiB was experimentally found to be enough to saturate the bus at
+// SuperSpeed+, so we first try double that for writes. If the operation fails
+// due to a lack of contiguous regions (or an ancient kernel), try smaller sizes
+// until we find one that works (see LinuxUsbTransport::Write). Reads are less
+// performance critical so for now just use a known good size.
+#define MAX_USBFS_BULK_WRITE_SIZE (256 * 1024)
+#define MAX_USBFS_BULK_READ_SIZE (16 * 1024)
+
+// This size should pretty much always work (it's compatible with pre-3.3
+// kernels and it's what we used to use historically), so if it doesn't work
+// something has gone badly wrong.
+#define MIN_USBFS_BULK_WRITE_SIZE (16 * 1024)
struct usb_handle
{
@@ -108,6 +119,7 @@
private:
std::unique_ptr<usb_handle> handle_;
const uint32_t ms_timeout_;
+ size_t max_usbfs_bulk_write_size_ = MAX_USBFS_BULK_WRITE_SIZE;
DISALLOW_COPY_AND_ASSIGN(LinuxUsbTransport);
};
@@ -269,6 +281,9 @@
auto path = android::base::StringPrintf("/sys/bus/usb/devices/%s/%s:1.%d/interface",
sysfs_name, sysfs_name, ifc->bInterfaceNumber);
if (android::base::ReadFileToString(path, &interface)) {
+ if (!interface.empty() && interface.back() == '\n') {
+ interface.pop_back();
+ }
snprintf(info.interface, sizeof(info.interface), "%s", interface.c_str());
}
@@ -404,34 +419,90 @@
{
unsigned char *data = (unsigned char*) _data;
unsigned count = 0;
- struct usbdevfs_bulktransfer bulk;
- int n;
+ struct usbdevfs_urb urb[2] = {};
+ bool pending[2] = {};
if (handle_->ep_out == 0 || handle_->desc == -1) {
return -1;
}
- do {
- int xfer;
- xfer = (len > MAX_USBFS_BULK_SIZE) ? MAX_USBFS_BULK_SIZE : len;
+ auto submit_urb = [&](size_t i) {
+ while (true) {
+ int xfer = (len > max_usbfs_bulk_write_size_) ? max_usbfs_bulk_write_size_ : len;
- bulk.ep = handle_->ep_out;
- bulk.len = xfer;
- bulk.data = data;
- bulk.timeout = ms_timeout_;
+ urb[i].type = USBDEVFS_URB_TYPE_BULK;
+ urb[i].endpoint = handle_->ep_out;
+ urb[i].buffer_length = xfer;
+ urb[i].buffer = data;
+ urb[i].usercontext = (void *)i;
- n = ioctl(handle_->desc, USBDEVFS_BULK, &bulk);
- if(n != xfer) {
- DBG("ERROR: n = %d, errno = %d (%s)\n",
- n, errno, strerror(errno));
+ int n = ioctl(handle_->desc, USBDEVFS_SUBMITURB, &urb[i]);
+ if (n != 0) {
+ if (errno == ENOMEM && max_usbfs_bulk_write_size_ > MIN_USBFS_BULK_WRITE_SIZE) {
+ max_usbfs_bulk_write_size_ /= 2;
+ continue;
+ }
+ DBG("ioctl(USBDEVFS_SUBMITURB) failed\n");
+ return false;
+ }
+
+ pending[i] = true;
+ count += xfer;
+ len -= xfer;
+ data += xfer;
+
+ return true;
+ }
+ };
+
+ auto reap_urb = [&](size_t i) {
+ while (pending[i]) {
+ struct usbdevfs_urb *urbp;
+ int res = ioctl(handle_->desc, USBDEVFS_REAPURB, &urbp);
+ if (res != 0) {
+ DBG("ioctl(USBDEVFS_REAPURB) failed\n");
+ return false;
+ }
+ size_t done = (size_t)urbp->usercontext;
+ if (done >= 2 || !pending[done]) {
+ DBG("unexpected urb\n");
+ return false;
+ }
+ if (urbp->status != 0 || urbp->actual_length != urbp->buffer_length) {
+ DBG("urb returned error\n");
+ return false;
+ }
+ pending[done] = false;
+ }
+ return true;
+ };
+
+ if (!submit_urb(0)) {
+ return -1;
+ }
+ while (len > 0) {
+ if (!submit_urb(1)) {
return -1;
}
-
- count += xfer;
- len -= xfer;
- data += xfer;
- } while(len > 0);
-
+ if (!reap_urb(0)) {
+ return -1;
+ }
+ if (len <= 0) {
+ if (!reap_urb(1)) {
+ return -1;
+ }
+ return count;
+ }
+ if (!submit_urb(0)) {
+ return -1;
+ }
+ if (!reap_urb(1)) {
+ return -1;
+ }
+ }
+ if (!reap_urb(0)) {
+ return -1;
+ }
return count;
}
@@ -447,7 +518,7 @@
}
while (len > 0) {
- int xfer = (len > MAX_USBFS_BULK_SIZE) ? MAX_USBFS_BULK_SIZE : len;
+ int xfer = (len > MAX_USBFS_BULK_READ_SIZE) ? MAX_USBFS_BULK_READ_SIZE : len;
bulk.ep = handle_->ep_in;
bulk.len = xfer;
diff --git a/fastboot/util.cpp b/fastboot/util.cpp
index e03012a..5966aea 100644
--- a/fastboot/util.cpp
+++ b/fastboot/util.cpp
@@ -31,7 +31,7 @@
#include <stdlib.h>
#include <string.h>
#include <sys/stat.h>
-#include <sys/time.h>
+#include <time.h>
#include <android-base/parseint.h>
#include <android-base/strings.h>
@@ -43,9 +43,9 @@
static bool g_verbose = false;
double now() {
- struct timeval tv;
- gettimeofday(&tv, NULL);
- return (double)tv.tv_sec + (double)tv.tv_usec / 1000000;
+ struct timespec ts;
+ clock_gettime(CLOCK_MONOTONIC, &ts);
+ return (double)ts.tv_sec + (double)ts.tv_nsec / 1000000000;
}
void die(const char* fmt, ...) {
diff --git a/fs_mgr/fs_mgr_remount.cpp b/fs_mgr/fs_mgr_remount.cpp
index 733ba2f..79c0b6d 100644
--- a/fs_mgr/fs_mgr_remount.cpp
+++ b/fs_mgr/fs_mgr_remount.cpp
@@ -42,6 +42,7 @@
#include <fstab/fstab.h>
#include <libavb_user/libavb_user.h>
#include <libgsi/libgsid.h>
+#include <private/android_filesystem_config.h>
#include "fs_mgr_overlayfs_control.h"
#include "fs_mgr_overlayfs_mount.h"
@@ -608,7 +609,19 @@
}
// Make sure we are root.
- if (::getuid() != 0) {
+ if (const auto uid = ::getuid(); uid != AID_ROOT) {
+ // If requesting auto reboot, also try to auto gain root.
+ if (auto_reboot && uid == AID_SHELL && access("/system/xbin/su", F_OK) == 0) {
+ std::vector<char*> args{const_cast<char*>("/system/xbin/su"),
+ const_cast<char*>("root")};
+ for (int i = 0; i < argc; ++i) {
+ args.push_back(argv[i]);
+ }
+ args.push_back(nullptr);
+ LOG(INFO) << "Attempting to gain root with \"su root\"";
+ execv(args[0], args.data());
+ PLOG(ERROR) << "Failed to execute \"su root\"";
+ }
LOG(ERROR) << "Not running as root. Try \"adb root\" first.";
return EXIT_FAILURE;
}
diff --git a/fs_mgr/libdm/dm_target.cpp b/fs_mgr/libdm/dm_target.cpp
index 90d91a0..1f6bd1a 100644
--- a/fs_mgr/libdm/dm_target.cpp
+++ b/fs_mgr/libdm/dm_target.cpp
@@ -61,6 +61,10 @@
return block_device_ + " " + std::to_string(physical_sector_);
}
+std::string DmTargetStripe::GetParameterString() const {
+ return "2 " + std::to_string(chunksize) + " " + block_device0_ + " 0 " + block_device1_ + " 0";
+}
+
DmTargetVerity::DmTargetVerity(uint64_t start, uint64_t length, uint32_t version,
const std::string& block_device, const std::string& hash_device,
uint32_t data_block_size, uint32_t hash_block_size,
diff --git a/fs_mgr/libdm/dm_test.cpp b/fs_mgr/libdm/dm_test.cpp
index a0129c2..d043be6 100644
--- a/fs_mgr/libdm/dm_test.cpp
+++ b/fs_mgr/libdm/dm_test.cpp
@@ -181,6 +181,13 @@
ASSERT_EQ(dm.GetState(dev.name()), DmDeviceState::ACTIVE);
}
+TEST_F(DmTest, StripeArgs) {
+ DmTargetStripe target(0, 4096, 1024, "/dev/loop0", "/dev/loop1");
+ ASSERT_EQ(target.name(), "striped");
+ ASSERT_TRUE(target.Valid());
+ ASSERT_EQ(target.GetParameterString(), "2 1024 /dev/loop0 0 /dev/loop1 0");
+}
+
TEST_F(DmTest, DmVerityArgsAvb2) {
std::string device = "/dev/block/platform/soc/1da4000.ufshc/by-name/vendor_a";
std::string algorithm = "sha1";
diff --git a/fs_mgr/libdm/include/libdm/dm_target.h b/fs_mgr/libdm/include/libdm/dm_target.h
index 09fe200..97f3c13 100644
--- a/fs_mgr/libdm/include/libdm/dm_target.h
+++ b/fs_mgr/libdm/include/libdm/dm_target.h
@@ -116,6 +116,24 @@
uint64_t physical_sector_;
};
+class DmTargetStripe final : public DmTarget {
+ public:
+ DmTargetStripe(uint64_t start, uint64_t length, uint64_t chunksize,
+ const std::string& block_device0, const std::string& block_device1)
+ : DmTarget(start, length),
+ chunksize(chunksize),
+ block_device0_(block_device0),
+ block_device1_(block_device1) {}
+
+ std::string name() const override { return "striped"; }
+ std::string GetParameterString() const override;
+
+ private:
+ uint64_t chunksize;
+ std::string block_device0_;
+ std::string block_device1_;
+};
+
class DmTargetVerity final : public DmTarget {
public:
DmTargetVerity(uint64_t start, uint64_t length, uint32_t version,
diff --git a/fs_mgr/libsnapshot/Android.bp b/fs_mgr/libsnapshot/Android.bp
index a8a7716..5ceaf28 100644
--- a/fs_mgr/libsnapshot/Android.bp
+++ b/fs_mgr/libsnapshot/Android.bp
@@ -108,7 +108,7 @@
],
srcs: [":libsnapshot_sources"],
static_libs: [
- "libfs_mgr_binder"
+ "libfs_mgr_binder",
],
}
@@ -128,12 +128,12 @@
static_libs: [
"libc++fs",
"libsnapshot_cow",
- ]
+ ],
}
cc_library_static {
name: "libsnapshot_init",
- native_coverage : true,
+ native_coverage: true,
defaults: ["libsnapshot_defaults"],
srcs: [":libsnapshot_sources"],
ramdisk_available: true,
@@ -204,6 +204,10 @@
"libsnapshot_cow/writer_v2.cpp",
"libsnapshot_cow/writer_v3.cpp",
],
+
+ header_libs: [
+ "libstorage_literals_headers",
+ ],
export_include_dirs: ["include"],
host_supported: true,
recovery_available: true,
@@ -243,7 +247,10 @@
cc_defaults {
name: "libsnapshot_test_defaults",
- defaults: ["libsnapshot_defaults", "libsnapshot_cow_defaults"],
+ defaults: [
+ "libsnapshot_defaults",
+ "libsnapshot_cow_defaults",
+ ],
srcs: [
"partition_cow_creator_test.cpp",
"snapshot_metadata_updater_test.cpp",
@@ -283,10 +290,13 @@
cc_test {
name: "vts_libsnapshot_test",
- defaults: ["libsnapshot_test_defaults", "libsnapshot_hal_deps"],
+ defaults: [
+ "libsnapshot_test_defaults",
+ "libsnapshot_hal_deps",
+ ],
test_suites: [
"vts",
- "device-tests"
+ "device-tests",
],
test_options: {
min_shipping_api_level: 30,
@@ -295,12 +305,15 @@
cc_test {
name: "vab_legacy_tests",
- defaults: ["libsnapshot_test_defaults", "libsnapshot_hal_deps"],
+ defaults: [
+ "libsnapshot_test_defaults",
+ "libsnapshot_hal_deps",
+ ],
cppflags: [
"-DLIBSNAPSHOT_TEST_VAB_LEGACY",
],
test_suites: [
- "device-tests"
+ "device-tests",
],
test_options: {
// Legacy VAB launched in Android R.
@@ -310,12 +323,15 @@
cc_test {
name: "vabc_legacy_tests",
- defaults: ["libsnapshot_test_defaults", "libsnapshot_hal_deps"],
+ defaults: [
+ "libsnapshot_test_defaults",
+ "libsnapshot_hal_deps",
+ ],
cppflags: [
"-DLIBSNAPSHOT_TEST_VABC_LEGACY",
],
test_suites: [
- "device-tests"
+ "device-tests",
],
test_options: {
// Legacy VABC launched in Android S.
@@ -343,7 +359,10 @@
cc_binary {
name: "snapshotctl",
- defaults: ["libsnapshot_cow_defaults", "libsnapshot_hal_deps"],
+ defaults: [
+ "libsnapshot_cow_defaults",
+ "libsnapshot_hal_deps",
+ ],
srcs: [
"snapshotctl.cpp",
],
@@ -412,8 +431,11 @@
"libgtest",
"libsnapshot_cow",
],
+ header_libs: [
+ "libstorage_literals_headers",
+ ],
test_suites: [
- "device-tests"
+ "device-tests",
],
test_options: {
min_shipping_api_level: 30,
diff --git a/fs_mgr/libsnapshot/include/libsnapshot/cow_compress.h b/fs_mgr/libsnapshot/include/libsnapshot/cow_compress.h
index 8191d61..ac04245 100644
--- a/fs_mgr/libsnapshot/include/libsnapshot/cow_compress.h
+++ b/fs_mgr/libsnapshot/include/libsnapshot/cow_compress.h
@@ -17,6 +17,7 @@
#pragma once
#include <memory>
+#include <vector>
#include "libsnapshot/cow_format.h"
namespace android {
@@ -40,8 +41,7 @@
uint32_t GetCompressionLevel() const { return compression_level_; }
uint32_t GetBlockSize() const { return block_size_; }
- [[nodiscard]] virtual std::basic_string<uint8_t> Compress(const void* data,
- size_t length) const = 0;
+ [[nodiscard]] virtual std::vector<uint8_t> Compress(const void* data, size_t length) const = 0;
private:
uint32_t compression_level_;
diff --git a/fs_mgr/libsnapshot/include/libsnapshot/cow_format.h b/fs_mgr/libsnapshot/include/libsnapshot/cow_format.h
index d410c14..6865b19 100644
--- a/fs_mgr/libsnapshot/include/libsnapshot/cow_format.h
+++ b/fs_mgr/libsnapshot/include/libsnapshot/cow_format.h
@@ -201,6 +201,12 @@
static constexpr uint64_t kCowOpSourceInfoTypeBit = 60;
static constexpr uint64_t kCowOpSourceInfoTypeNumBits = 4;
static constexpr uint64_t kCowOpSourceInfoTypeMask = (1ULL << kCowOpSourceInfoTypeNumBits) - 1;
+
+static constexpr uint64_t kCowOpSourceInfoCompressionBit = 57;
+static constexpr uint64_t kCowOpSourceInfoCompressionNumBits = 3;
+static constexpr uint64_t kCowOpSourceInfoCompressionMask =
+ ((1ULL << kCowOpSourceInfoCompressionNumBits) - 1);
+
// The on disk format of cow (currently == CowOperation)
struct CowOperationV3 {
// If this operation reads from the data section of the COW, this contains
@@ -211,8 +217,8 @@
uint32_t new_block;
// source_info with have the following layout
- // |---4 bits ---| ---12 bits---| --- 48 bits ---|
- // |--- type --- | -- unused -- | --- source --- |
+ // |--- 4 bits -- | --------- 3 bits ------ | --- 9 bits --- | --- 48 bits ---|
+ // |--- type --- | -- compression factor --| --- unused --- | --- source --- |
//
// The value of |source| depends on the operation code.
//
@@ -225,6 +231,17 @@
// For ops other than Label:
// Bits 47-62 are reserved and must be zero.
// A block is compressed if it’s data is < block_sz
+ //
+ // Bits [57-59] represents the compression factor.
+ //
+ // Compression - factor
+ // ==========================
+ // 000 - 4k
+ // 001 - 8k
+ // 010 - 16k
+ // ...
+ // 110 - 256k
+ //
uint64_t source_info_;
constexpr uint64_t source() const { return source_info_ & kCowOpSourceInfoDataMask; }
constexpr void set_source(uint64_t source) {
@@ -245,6 +262,20 @@
source_info_ |= (static_cast<uint64_t>(type) & kCowOpSourceInfoTypeMask)
<< kCowOpSourceInfoTypeBit;
}
+ constexpr void set_compression_bits(uint8_t compression_factor) {
+ // Clear the 3 bits from bit 57 - [57-59]
+ source_info_ &= ~(kCowOpSourceInfoCompressionMask << kCowOpSourceInfoCompressionBit);
+ // Set the actual compression factor
+ source_info_ |=
+ (static_cast<uint64_t>(compression_factor) & kCowOpSourceInfoCompressionMask)
+ << kCowOpSourceInfoCompressionBit;
+ }
+ constexpr uint8_t compression_bits() const {
+ // Grab the 3 bits from [57-59]
+ const auto compression_factor =
+ (source_info_ >> kCowOpSourceInfoCompressionBit) & kCowOpSourceInfoCompressionMask;
+ return static_cast<uint8_t>(compression_factor);
+ }
} __attribute__((packed));
// Ensure that getters/setters added to CowOperationV3 does not increases size
@@ -326,5 +357,11 @@
// Convert compression name to internal value.
std::optional<CowCompressionAlgorithm> CompressionAlgorithmFromString(std::string_view name);
+// Return block size used for compression
+size_t CowOpCompressionSize(const CowOperation* op, size_t block_size);
+
+// Return the relative offset of the I/O block which the CowOperation
+// multi-block compression
+bool GetBlockOffset(const CowOperation* op, uint64_t io_block, size_t block_size, off_t* offset);
} // namespace snapshot
} // namespace android
diff --git a/fs_mgr/libsnapshot/include/libsnapshot/cow_reader.h b/fs_mgr/libsnapshot/include/libsnapshot/cow_reader.h
index bf4c79f..3f49c69 100644
--- a/fs_mgr/libsnapshot/include/libsnapshot/cow_reader.h
+++ b/fs_mgr/libsnapshot/include/libsnapshot/cow_reader.h
@@ -162,6 +162,9 @@
// Creates a clone of the current CowReader without the file handlers
std::unique_ptr<CowReader> CloneCowReader();
+ // Get the max compression size
+ uint32_t GetMaxCompressionSize();
+
void UpdateMergeOpsCompleted(int num_merge_ops) { header_.num_merge_ops += num_merge_ops; }
private:
diff --git a/fs_mgr/libsnapshot/include/libsnapshot/cow_writer.h b/fs_mgr/libsnapshot/include/libsnapshot/cow_writer.h
index 0194ffd..651083f 100644
--- a/fs_mgr/libsnapshot/include/libsnapshot/cow_writer.h
+++ b/fs_mgr/libsnapshot/include/libsnapshot/cow_writer.h
@@ -119,36 +119,38 @@
class CompressWorker {
public:
- CompressWorker(std::unique_ptr<ICompressor>&& compressor, uint32_t block_size);
+ CompressWorker(std::unique_ptr<ICompressor>&& compressor);
bool RunThread();
- void EnqueueCompressBlocks(const void* buffer, size_t num_blocks);
- bool GetCompressedBuffers(std::vector<std::basic_string<uint8_t>>* compressed_buf);
+ void EnqueueCompressBlocks(const void* buffer, size_t block_size, size_t num_blocks);
+ bool GetCompressedBuffers(std::vector<std::vector<uint8_t>>* compressed_buf);
void Finalize();
static uint32_t GetDefaultCompressionLevel(CowCompressionAlgorithm compression);
static bool CompressBlocks(ICompressor* compressor, size_t block_size, const void* buffer,
size_t num_blocks,
- std::vector<std::basic_string<uint8_t>>* compressed_data);
+ std::vector<std::vector<uint8_t>>* compressed_data);
private:
struct CompressWork {
const void* buffer;
size_t num_blocks;
+ size_t block_size;
bool compression_status = false;
- std::vector<std::basic_string<uint8_t>> compressed_data;
+ std::vector<std::vector<uint8_t>> compressed_data;
};
std::unique_ptr<ICompressor> compressor_;
- uint32_t block_size_;
std::queue<CompressWork> work_queue_;
std::queue<CompressWork> compressed_queue_;
std::mutex lock_;
std::condition_variable cv_;
bool stopped_ = false;
+ size_t total_submitted_ = 0;
+ size_t total_processed_ = 0;
- bool CompressBlocks(const void* buffer, size_t num_blocks,
- std::vector<std::basic_string<uint8_t>>* compressed_data);
+ bool CompressBlocks(const void* buffer, size_t num_blocks, size_t block_size,
+ std::vector<std::vector<uint8_t>>* compressed_data);
};
// Create an ICowWriter not backed by any file. This is useful for estimating
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/cow_compress.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/cow_compress.cpp
index abc7d33..0205f50 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/cow_compress.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/cow_compress.cpp
@@ -20,6 +20,7 @@
#include <limits>
#include <memory>
#include <queue>
+#include <vector>
#include <android-base/file.h>
#include <android-base/logging.h>
@@ -103,9 +104,9 @@
GzCompressor(uint32_t compression_level, const uint32_t block_size)
: ICompressor(compression_level, block_size){};
- std::basic_string<uint8_t> Compress(const void* data, size_t length) const override {
+ std::vector<uint8_t> Compress(const void* data, size_t length) const override {
const auto bound = compressBound(length);
- std::basic_string<uint8_t> buffer(bound, '\0');
+ std::vector<uint8_t> buffer(bound, '\0');
uLongf dest_len = bound;
auto rv = compress2(buffer.data(), &dest_len, reinterpret_cast<const Bytef*>(data), length,
@@ -124,13 +125,13 @@
Lz4Compressor(uint32_t compression_level, const uint32_t block_size)
: ICompressor(compression_level, block_size){};
- std::basic_string<uint8_t> Compress(const void* data, size_t length) const override {
+ std::vector<uint8_t> Compress(const void* data, size_t length) const override {
const auto bound = LZ4_compressBound(length);
if (!bound) {
LOG(ERROR) << "LZ4_compressBound returned 0";
return {};
}
- std::basic_string<uint8_t> buffer(bound, '\0');
+ std::vector<uint8_t> buffer(bound, '\0');
const auto compressed_size =
LZ4_compress_default(static_cast<const char*>(data),
@@ -156,13 +157,13 @@
BrotliCompressor(uint32_t compression_level, const uint32_t block_size)
: ICompressor(compression_level, block_size){};
- std::basic_string<uint8_t> Compress(const void* data, size_t length) const override {
+ std::vector<uint8_t> Compress(const void* data, size_t length) const override {
const auto bound = BrotliEncoderMaxCompressedSize(length);
if (!bound) {
LOG(ERROR) << "BrotliEncoderMaxCompressedSize returned 0";
return {};
}
- std::basic_string<uint8_t> buffer(bound, '\0');
+ std::vector<uint8_t> buffer(bound, '\0');
size_t encoded_size = bound;
auto rv = BrotliEncoderCompress(
@@ -186,8 +187,8 @@
ZSTD_CCtx_setParameter(zstd_context_.get(), ZSTD_c_windowLog, log2(GetBlockSize()));
};
- std::basic_string<uint8_t> Compress(const void* data, size_t length) const override {
- std::basic_string<uint8_t> buffer(ZSTD_compressBound(length), '\0');
+ std::vector<uint8_t> Compress(const void* data, size_t length) const override {
+ std::vector<uint8_t> buffer(ZSTD_compressBound(length), '\0');
const auto compressed_size =
ZSTD_compress2(zstd_context_.get(), buffer.data(), buffer.size(), data, length);
if (compressed_size <= 0) {
@@ -208,14 +209,14 @@
std::unique_ptr<ZSTD_CCtx, decltype(&ZSTD_freeCCtx)> zstd_context_;
};
-bool CompressWorker::CompressBlocks(const void* buffer, size_t num_blocks,
- std::vector<std::basic_string<uint8_t>>* compressed_data) {
- return CompressBlocks(compressor_.get(), block_size_, buffer, num_blocks, compressed_data);
+bool CompressWorker::CompressBlocks(const void* buffer, size_t num_blocks, size_t block_size,
+ std::vector<std::vector<uint8_t>>* compressed_data) {
+ return CompressBlocks(compressor_.get(), block_size, buffer, num_blocks, compressed_data);
}
bool CompressWorker::CompressBlocks(ICompressor* compressor, size_t block_size, const void* buffer,
size_t num_blocks,
- std::vector<std::basic_string<uint8_t>>* compressed_data) {
+ std::vector<std::vector<uint8_t>>* compressed_data) {
const uint8_t* iter = reinterpret_cast<const uint8_t*>(buffer);
while (num_blocks) {
auto data = compressor->Compress(iter, block_size);
@@ -223,7 +224,7 @@
PLOG(ERROR) << "CompressBlocks: Compression failed";
return false;
}
- if (data.size() > std::numeric_limits<uint16_t>::max()) {
+ if (data.size() > std::numeric_limits<uint32_t>::max()) {
LOG(ERROR) << "Compressed block is too large: " << data.size();
return false;
}
@@ -254,7 +255,8 @@
}
// Compress blocks
- bool ret = CompressBlocks(blocks.buffer, blocks.num_blocks, &blocks.compressed_data);
+ bool ret = CompressBlocks(blocks.buffer, blocks.num_blocks, blocks.block_size,
+ &blocks.compressed_data);
blocks.compression_status = ret;
{
std::lock_guard<std::mutex> lock(lock_);
@@ -273,35 +275,31 @@
return true;
}
-void CompressWorker::EnqueueCompressBlocks(const void* buffer, size_t num_blocks) {
+void CompressWorker::EnqueueCompressBlocks(const void* buffer, size_t block_size,
+ size_t num_blocks) {
{
std::lock_guard<std::mutex> lock(lock_);
CompressWork blocks = {};
blocks.buffer = buffer;
+ blocks.block_size = block_size;
blocks.num_blocks = num_blocks;
work_queue_.push(std::move(blocks));
+ total_submitted_ += 1;
}
cv_.notify_all();
}
-bool CompressWorker::GetCompressedBuffers(std::vector<std::basic_string<uint8_t>>* compressed_buf) {
- {
+bool CompressWorker::GetCompressedBuffers(std::vector<std::vector<uint8_t>>* compressed_buf) {
+ while (true) {
std::unique_lock<std::mutex> lock(lock_);
- while (compressed_queue_.empty() && !stopped_) {
+ while ((total_submitted_ != total_processed_) && compressed_queue_.empty() && !stopped_) {
cv_.wait(lock);
}
-
- if (stopped_) {
- return true;
- }
- }
-
- {
- std::lock_guard<std::mutex> lock(lock_);
while (compressed_queue_.size() > 0) {
CompressWork blocks = std::move(compressed_queue_.front());
compressed_queue_.pop();
+ total_processed_ += 1;
if (blocks.compression_status) {
compressed_buf->insert(compressed_buf->end(),
@@ -312,9 +310,12 @@
return false;
}
}
+ if ((total_submitted_ == total_processed_) || stopped_) {
+ total_submitted_ = 0;
+ total_processed_ = 0;
+ return true;
+ }
}
-
- return true;
}
std::unique_ptr<ICompressor> ICompressor::Brotli(uint32_t compression_level,
@@ -344,8 +345,8 @@
cv_.notify_all();
}
-CompressWorker::CompressWorker(std::unique_ptr<ICompressor>&& compressor, uint32_t block_size)
- : compressor_(std::move(compressor)), block_size_(block_size) {}
+CompressWorker::CompressWorker(std::unique_ptr<ICompressor>&& compressor)
+ : compressor_(std::move(compressor)) {}
} // namespace snapshot
} // namespace android
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/cow_format.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/cow_format.cpp
index 8d1786c..19014c0 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/cow_format.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/cow_format.cpp
@@ -21,6 +21,7 @@
#include <android-base/logging.h>
#include <android-base/stringprintf.h>
#include <libsnapshot/cow_format.h>
+#include <storage_literals/storage_literals.h>
#include "writer_v2.h"
#include "writer_v3.h"
@@ -28,6 +29,7 @@
namespace snapshot {
using android::base::unique_fd;
+using namespace android::storage_literals;
std::ostream& EmitCowTypeString(std::ostream& os, CowOperationType cow_type) {
switch (cow_type) {
@@ -174,5 +176,36 @@
return CreateCowWriter(version, options, unique_fd{-1}, std::nullopt);
}
+size_t CowOpCompressionSize(const CowOperation* op, size_t block_size) {
+ uint8_t compression_bits = op->compression_bits();
+ return (block_size << compression_bits);
+}
+
+bool GetBlockOffset(const CowOperation* op, uint64_t io_block, size_t block_size, off_t* offset) {
+ const uint64_t new_block = op->new_block;
+
+ if (op->type() != kCowReplaceOp || io_block < new_block) {
+ LOG(VERBOSE) << "Invalid IO request for block: " << io_block
+ << " CowOperation: new_block: " << new_block;
+ return false;
+ }
+
+ // Get the actual compression size
+ const size_t compression_size = CowOpCompressionSize(op, block_size);
+ // Find the number of blocks spanned
+ const size_t num_blocks = compression_size / block_size;
+ // Find the distance of the I/O block which this
+ // CowOperation encompasses
+ const size_t block_distance = io_block - new_block;
+ // Check if this block is within this range;
+ // if so, return the relative offset
+ if (block_distance < num_blocks) {
+ *offset = block_distance * block_size;
+ return true;
+ }
+
+ return false;
+}
+
} // namespace snapshot
} // namespace android
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/cow_reader.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/cow_reader.cpp
index 1b4a971..6516499 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/cow_reader.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/cow_reader.cpp
@@ -26,6 +26,7 @@
#include <android-base/logging.h>
#include <libsnapshot/cow_format.h>
#include <libsnapshot/cow_reader.h>
+#include <storage_literals/storage_literals.h>
#include <zlib.h>
#include "cow_decompress.h"
@@ -35,6 +36,8 @@
namespace android {
namespace snapshot {
+using namespace android::storage_literals;
+
bool ReadCowHeader(android::base::borrowed_fd fd, CowHeaderV3* header) {
if (lseek(fd.get(), 0, SEEK_SET) < 0) {
PLOG(ERROR) << "lseek header failed";
@@ -161,6 +164,21 @@
return PrepMergeOps();
}
+uint32_t CowReader::GetMaxCompressionSize() {
+ switch (header_.prefix.major_version) {
+ case 1:
+ case 2:
+ // Old versions supports only 4KB compression.
+ return header_.block_size;
+ ;
+ case 3:
+ return header_.max_compression_size;
+ default:
+ LOG(ERROR) << "Unknown version: " << header_.prefix.major_version;
+ return 0;
+ }
+}
+
//
// This sets up the data needed for MergeOpIter. MergeOpIter presents
// data in the order we intend to merge in.
@@ -705,6 +723,11 @@
ssize_t CowReader::ReadData(const CowOperation* op, void* buffer, size_t buffer_size,
size_t ignore_bytes) {
std::unique_ptr<IDecompressor> decompressor;
+ const size_t op_buf_size = CowOpCompressionSize(op, header_.block_size);
+ if (!op_buf_size) {
+ LOG(ERROR) << "Compression size is zero. op: " << *op;
+ return -1;
+ }
switch (GetCompressionType()) {
case kCowCompressNone:
break;
@@ -715,12 +738,12 @@
decompressor = IDecompressor::Brotli();
break;
case kCowCompressZstd:
- if (header_.block_size != op->data_length) {
+ if (op_buf_size != op->data_length) {
decompressor = IDecompressor::Zstd();
}
break;
case kCowCompressLz4:
- if (header_.block_size != op->data_length) {
+ if (op_buf_size != op->data_length) {
decompressor = IDecompressor::Lz4();
}
break;
@@ -736,14 +759,14 @@
offset = op->source();
}
if (!decompressor ||
- ((op->data_length == header_.block_size) && (header_.prefix.major_version == 3))) {
+ ((op->data_length == op_buf_size) && (header_.prefix.major_version == 3))) {
CowDataStream stream(this, offset + ignore_bytes, op->data_length - ignore_bytes);
return stream.ReadFully(buffer, buffer_size);
}
CowDataStream stream(this, offset, op->data_length);
decompressor->set_stream(&stream);
- return decompressor->Decompress(buffer, buffer_size, header_.block_size, ignore_bytes);
+ return decompressor->Decompress(buffer, buffer_size, op_buf_size, ignore_bytes);
}
bool CowReader::GetSourceOffset(const CowOperation* op, uint64_t* source_offset) {
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/create_cow.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/create_cow.cpp
index efb1035..5497b72 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/create_cow.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/create_cow.cpp
@@ -71,6 +71,8 @@
const int kNumThreads = 6;
const size_t kBlockSizeToRead = 1_MiB;
+ const size_t compression_factor_ = 64_KiB;
+ size_t replace_ops_ = 0, copy_ops_ = 0, zero_ops_ = 0, in_place_ops_ = 0;
std::unordered_map<std::string, int> source_block_hash_;
std::mutex source_block_hash_lock_;
@@ -81,7 +83,12 @@
std::unique_ptr<uint8_t[]> zblock_;
std::string compression_ = "lz4";
- unique_fd fd_;
+ unique_fd cow_fd_;
+ unique_fd target_fd_;
+
+ std::vector<uint64_t> zero_blocks_;
+ std::vector<uint64_t> replace_blocks_;
+ std::unordered_map<uint64_t, uint64_t> copy_blocks_;
const int BLOCK_SZ = 4_KiB;
void SHA256(const void* data, size_t length, uint8_t out[32]);
@@ -93,7 +100,14 @@
bool FindSourceBlockHash();
bool PrepareParse(std::string& parsing_file, const bool createSnapshot);
bool ParsePartition();
- bool WriteSnapshot(const void* buffer, uint64_t block, std::string& block_hash);
+ void PrepareMergeBlock(const void* buffer, uint64_t block, std::string& block_hash);
+ bool WriteV3Snapshots();
+ size_t PrepareWrite(size_t* pending_ops, size_t start_index);
+
+ bool CreateSnapshotWriter();
+ bool WriteOrderedSnapshots();
+ bool WriteNonOrderedSnapshots();
+ bool VerifyMergeOrder();
};
void CreateSnapshotLogger(android::base::LogId, android::base::LogSeverity severity, const char*,
@@ -118,21 +132,19 @@
create_snapshot_patch_ = createSnapshot;
if (createSnapshot) {
- fd_.reset(open(patch_file_.c_str(), O_RDWR | O_CREAT | O_TRUNC, 0666));
- if (fd_ < 0) {
+ cow_fd_.reset(open(patch_file_.c_str(), O_RDWR | O_CREAT | O_TRUNC, 0666));
+ if (cow_fd_ < 0) {
PLOG(ERROR) << "Failed to open the snapshot-patch file: " << patch_file_;
return false;
}
+ target_fd_.reset((open(parsing_file_.c_str(), O_RDONLY)));
+ if (target_fd_ < 0) {
+ LOG(ERROR) << "open failed: " << parsing_file_;
+ return false;
+ }
zblock_ = std::make_unique<uint8_t[]>(BLOCK_SZ);
std::memset(zblock_.get(), 0, BLOCK_SZ);
-
- CowOptions options;
- options.compression = compression_;
- options.num_compress_threads = 2;
- options.batch_write = true;
- options.cluster_ops = 600;
- writer_ = CreateCowWriter(2, options, std::move(fd_));
}
return true;
}
@@ -187,19 +199,158 @@
return out;
}
-bool CreateSnapshot::WriteSnapshot(const void* buffer, uint64_t block, std::string& block_hash) {
+void CreateSnapshot::PrepareMergeBlock(const void* buffer, uint64_t block,
+ std::string& block_hash) {
if (std::memcmp(zblock_.get(), buffer, BLOCK_SZ) == 0) {
std::lock_guard<std::mutex> lock(write_lock_);
- return writer_->AddZeroBlocks(block, 1);
+ zero_blocks_.push_back(block);
+ return;
}
auto iter = source_block_hash_.find(block_hash);
if (iter != source_block_hash_.end()) {
std::lock_guard<std::mutex> lock(write_lock_);
- return writer_->AddCopy(block, iter->second, 1);
+ // In-place copy is skipped
+ if (block != iter->second) {
+ copy_blocks_[block] = iter->second;
+ } else {
+ in_place_ops_ += 1;
+ }
+ return;
}
std::lock_guard<std::mutex> lock(write_lock_);
- return writer_->AddRawBlocks(block, buffer, BLOCK_SZ);
+ replace_blocks_.push_back(block);
+}
+
+size_t CreateSnapshot::PrepareWrite(size_t* pending_ops, size_t start_index) {
+ size_t num_ops = *pending_ops;
+ uint64_t start_block = replace_blocks_[start_index];
+ size_t nr_consecutive = 1;
+ num_ops -= 1;
+ while (num_ops) {
+ uint64_t next_block = replace_blocks_[start_index + nr_consecutive];
+ if (next_block != start_block + nr_consecutive) {
+ break;
+ }
+ nr_consecutive += 1;
+ num_ops -= 1;
+ }
+ return nr_consecutive;
+}
+
+bool CreateSnapshot::CreateSnapshotWriter() {
+ uint64_t dev_sz = lseek(target_fd_.get(), 0, SEEK_END);
+ CowOptions options;
+ options.compression = compression_;
+ options.num_compress_threads = 2;
+ options.batch_write = true;
+ options.cluster_ops = 600;
+ options.compression_factor = compression_factor_;
+ options.max_blocks = {dev_sz / options.block_size};
+ writer_ = CreateCowWriter(3, options, std::move(cow_fd_));
+ return true;
+}
+
+bool CreateSnapshot::WriteNonOrderedSnapshots() {
+ zero_ops_ = zero_blocks_.size();
+ for (auto it = zero_blocks_.begin(); it != zero_blocks_.end(); it++) {
+ if (!writer_->AddZeroBlocks(*it, 1)) {
+ return false;
+ }
+ }
+ std::string buffer(compression_factor_, '\0');
+
+ replace_ops_ = replace_blocks_.size();
+ size_t blocks_to_compress = replace_blocks_.size();
+ size_t num_ops = 0;
+ size_t block_index = 0;
+ while (blocks_to_compress) {
+ num_ops = std::min((compression_factor_ / BLOCK_SZ), blocks_to_compress);
+ auto linear_blocks = PrepareWrite(&num_ops, block_index);
+ if (!android::base::ReadFullyAtOffset(target_fd_.get(), buffer.data(),
+ (linear_blocks * BLOCK_SZ),
+ replace_blocks_[block_index] * BLOCK_SZ)) {
+ LOG(ERROR) << "Failed to read at offset: " << replace_blocks_[block_index] * BLOCK_SZ
+ << " size: " << linear_blocks * BLOCK_SZ;
+ return false;
+ }
+ if (!writer_->AddRawBlocks(replace_blocks_[block_index], buffer.data(),
+ linear_blocks * BLOCK_SZ)) {
+ LOG(ERROR) << "AddRawBlocks failed";
+ return false;
+ }
+
+ block_index += linear_blocks;
+ blocks_to_compress -= linear_blocks;
+ }
+ if (!writer_->Finalize()) {
+ return false;
+ }
+ return true;
+}
+
+bool CreateSnapshot::WriteOrderedSnapshots() {
+ std::unordered_map<uint64_t, uint64_t> overwritten_blocks;
+ std::vector<std::pair<uint64_t, uint64_t>> merge_sequence;
+ for (auto it = copy_blocks_.begin(); it != copy_blocks_.end(); it++) {
+ if (overwritten_blocks.count(it->second)) {
+ replace_blocks_.push_back(it->first);
+ continue;
+ }
+ overwritten_blocks[it->first] = it->second;
+ merge_sequence.emplace_back(std::make_pair(it->first, it->second));
+ }
+ // Sort the blocks so that if the blocks are contiguous, it would help
+ // compress multiple blocks in one shot based on the compression factor.
+ std::sort(replace_blocks_.begin(), replace_blocks_.end());
+
+ copy_ops_ = merge_sequence.size();
+ for (auto it = merge_sequence.begin(); it != merge_sequence.end(); it++) {
+ if (!writer_->AddCopy(it->first, it->second, 1)) {
+ return false;
+ }
+ }
+
+ return true;
+}
+
+bool CreateSnapshot::VerifyMergeOrder() {
+ unique_fd read_fd;
+ read_fd.reset(open(patch_file_.c_str(), O_RDONLY));
+ if (read_fd < 0) {
+ PLOG(ERROR) << "Failed to open the snapshot-patch file: " << patch_file_;
+ return false;
+ }
+ CowReader reader;
+ if (!reader.Parse(read_fd)) {
+ LOG(ERROR) << "Parse failed";
+ return false;
+ }
+
+ if (!reader.VerifyMergeOps()) {
+ LOG(ERROR) << "MergeOps Order is wrong";
+ return false;
+ }
+ return true;
+}
+
+bool CreateSnapshot::WriteV3Snapshots() {
+ if (!CreateSnapshotWriter()) {
+ return false;
+ }
+ if (!WriteOrderedSnapshots()) {
+ return false;
+ }
+ if (!WriteNonOrderedSnapshots()) {
+ return false;
+ }
+ if (!VerifyMergeOrder()) {
+ return false;
+ }
+
+ LOG(INFO) << "In-place: " << in_place_ops_ << " Zero: " << zero_ops_
+ << " Replace: " << replace_ops_ << " copy: " << copy_ops_;
+ return true;
}
bool CreateSnapshot::ReadBlocks(off_t offset, const int skip_blocks, const uint64_t dev_sz) {
@@ -241,10 +392,7 @@
std::string hash = ToHexString(checksum, sizeof(checksum));
if (create_snapshot_patch_) {
- if (!WriteSnapshot(bufptr, blkindex, hash)) {
- LOG(ERROR) << "WriteSnapshot failed for block: " << blkindex;
- return false;
- }
+ PrepareMergeBlock(bufptr, blkindex, hash);
} else {
std::lock_guard<std::mutex> lock(source_block_hash_lock_);
{
@@ -306,8 +454,8 @@
ret = t.get() && ret;
}
- if (ret && create_snapshot_patch_ && !writer_->Finalize()) {
- LOG(ERROR) << "Finzalize failed";
+ if (ret && create_snapshot_patch_ && !WriteV3Snapshots()) {
+ LOG(ERROR) << "Snapshot Write failed";
return false;
}
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/parser_v2.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/parser_v2.cpp
index fe977b7..a35b614 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/parser_v2.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/parser_v2.cpp
@@ -206,6 +206,8 @@
auto& new_op = out->ops->at(i);
new_op.set_type(v2_op.type);
+ // v2 ops always have 4k compression
+ new_op.set_compression_bits(0);
new_op.data_length = v2_op.data_length;
if (v2_op.new_block > std::numeric_limits<uint32_t>::max()) {
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/snapshot_reader.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/snapshot_reader.cpp
index 4e90a0f..12073fc 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/snapshot_reader.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/snapshot_reader.cpp
@@ -42,10 +42,21 @@
op_iter_->Next();
continue;
}
- if (op->new_block >= ops_.size()) {
- ops_.resize(op->new_block + 1, nullptr);
+
+ size_t num_blocks = 1;
+ if (op->type() == kCowReplaceOp) {
+ num_blocks = (CowOpCompressionSize(op, block_size_) / block_size_);
}
- ops_[op->new_block] = op;
+ if (op->new_block >= ops_.size()) {
+ ops_.resize(op->new_block + num_blocks, nullptr);
+ }
+
+ size_t vec_index = op->new_block;
+ while (num_blocks) {
+ ops_[vec_index] = op;
+ num_blocks -= 1;
+ vec_index += 1;
+ }
op_iter_->Next();
}
}
@@ -172,11 +183,20 @@
} else if (op->type() == kCowZeroOp) {
memset(buffer, 0, bytes_to_read);
} else if (op->type() == kCowReplaceOp) {
- if (cow_->ReadData(op, buffer, bytes_to_read, start_offset) < bytes_to_read) {
- LOG(ERROR) << "CompressedSnapshotReader failed to read replace op";
+ size_t buffer_size = CowOpCompressionSize(op, block_size_);
+ uint8_t temp_buffer[buffer_size];
+ if (cow_->ReadData(op, temp_buffer, buffer_size, 0) < buffer_size) {
+ LOG(ERROR) << "CompressedSnapshotReader failed to read replace op: buffer_size: "
+ << buffer_size << "start_offset: " << start_offset;
errno = EIO;
return -1;
}
+ off_t block_offset{};
+ if (!GetBlockOffset(op, chunk, block_size_, &block_offset)) {
+ LOG(ERROR) << "GetBlockOffset failed";
+ return -1;
+ }
+ std::memcpy(buffer, (char*)temp_buffer + block_offset + start_offset, bytes_to_read);
} else if (op->type() == kCowXorOp) {
borrowed_fd fd = GetSourceFd();
if (fd < 0) {
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/test_v2.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/test_v2.cpp
index 1d1d24c..ce80cd7 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/test_v2.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/test_v2.cpp
@@ -18,6 +18,7 @@
#include <iostream>
#include <memory>
#include <string_view>
+#include <vector>
#include <android-base/file.h>
#include <android-base/logging.h>
@@ -429,7 +430,7 @@
template <typename T>
class HorribleStream : public IByteStream {
public:
- HorribleStream(const std::basic_string<T>& input) : input_(input) {}
+ HorribleStream(const std::vector<T>& input) : input_(input) {}
ssize_t Read(void* buffer, size_t length) override {
if (pos_ >= input_.size()) {
@@ -444,16 +445,17 @@
size_t Size() const override { return input_.size(); }
private:
- std::basic_string<T> input_;
+ std::vector<T> input_;
size_t pos_ = 0;
};
TEST(HorribleStream, ReadFully) {
- std::string expected = "this is some data";
+ std::string expected_str = "this is some data";
+ std::vector<char> expected{expected_str.begin(), expected_str.end()};
HorribleStream<char> stream(expected);
- std::string buffer(expected.size(), '\0');
+ std::vector<char> buffer(expected.size(), '\0');
ASSERT_TRUE(stream.ReadFully(buffer.data(), buffer.size()));
ASSERT_EQ(buffer, expected);
}
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/test_v3.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/test_v3.cpp
index de60213..2021348 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/test_v3.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/test_v3.cpp
@@ -18,6 +18,7 @@
#include <libsnapshot/cow_format.h>
#include <libsnapshot/cow_reader.h>
#include <libsnapshot/cow_writer.h>
+#include <storage_literals/storage_literals.h>
#include "writer_v2.h"
#include "writer_v3.h"
@@ -29,6 +30,9 @@
namespace android {
namespace snapshot {
+using namespace android::storage_literals;
+using ::testing::TestWithParam;
+
class CowTestV3 : public ::testing::Test {
protected:
virtual void SetUp() override {
@@ -105,6 +109,40 @@
ASSERT_EQ(reader.header_v3().op_count, 20);
}
+TEST_F(CowTestV3, MaxOpSingleThreadCompression) {
+ CowOptions options;
+ options.op_count_max = 20;
+ options.num_compress_threads = 1;
+ options.compression_factor = 4096 * 8;
+ options.compression = "lz4";
+
+ auto writer = CreateCowWriter(3, options, GetCowFd());
+ ASSERT_TRUE(writer->AddZeroBlocks(1, 20));
+ std::string data = "This is some data, believe it";
+ data.resize(options.block_size, '\0');
+
+ ASSERT_FALSE(writer->AddRawBlocks(5, data.data(), data.size()));
+
+ ASSERT_TRUE(writer->Finalize());
+}
+
+TEST_F(CowTestV3, MaxOpMultiThreadCompression) {
+ CowOptions options;
+ options.op_count_max = 20;
+ options.num_compress_threads = 2;
+ options.compression_factor = 4096 * 8;
+ options.compression = "lz4";
+
+ auto writer = CreateCowWriter(3, options, GetCowFd());
+ ASSERT_TRUE(writer->AddZeroBlocks(1, 20));
+ std::string data = "This is some data, believe it";
+ data.resize(options.block_size, '\0');
+
+ ASSERT_FALSE(writer->AddRawBlocks(5, data.data(), data.size()));
+
+ ASSERT_TRUE(writer->Finalize());
+}
+
TEST_F(CowTestV3, ZeroOp) {
CowOptions options;
options.op_count_max = 20;
@@ -484,14 +522,14 @@
ASSERT_TRUE(reader.Parse(cow_->fd));
auto header = reader.header_v3();
- ASSERT_EQ(header.sequence_data_count, 0);
+ ASSERT_EQ(header.sequence_data_count, static_cast<uint64_t>(0));
ASSERT_EQ(header.resume_point_count, 0);
ASSERT_EQ(header.resume_point_max, 4);
writer->AddLabel(0);
ASSERT_TRUE(reader.Parse(cow_->fd));
header = reader.header_v3();
- ASSERT_EQ(header.sequence_data_count, 0);
+ ASSERT_EQ(header.sequence_data_count, static_cast<uint64_t>(0));
ASSERT_EQ(header.resume_point_count, 1);
ASSERT_EQ(header.resume_point_max, 4);
@@ -699,5 +737,189 @@
ASSERT_FALSE(writer->AddZeroBlocks(0, 19));
}
+struct TestParam {
+ std::string compression;
+ int block_size;
+ int num_threads;
+ size_t cluster_ops;
+};
+
+class VariableBlockTest : public ::testing::TestWithParam<TestParam> {
+ protected:
+ virtual void SetUp() override {
+ cow_ = std::make_unique<TemporaryFile>();
+ ASSERT_GE(cow_->fd, 0) << strerror(errno);
+ }
+
+ virtual void TearDown() override { cow_ = nullptr; }
+
+ unique_fd GetCowFd() { return unique_fd{dup(cow_->fd)}; }
+
+ std::unique_ptr<TemporaryFile> cow_;
+};
+
+// Helper to check read sizes.
+static inline void ReadBlockData(CowReader& reader, const CowOperation* op, void* buffer,
+ size_t size) {
+ size_t block_size = CowOpCompressionSize(op, 4096);
+ std::string data(block_size, '\0');
+ size_t value = reader.ReadData(op, data.data(), block_size);
+ ASSERT_TRUE(value == block_size);
+ std::memcpy(buffer, data.data(), size);
+}
+
+TEST_P(VariableBlockTest, VariableBlockCompressionTest) {
+ const TestParam params = GetParam();
+
+ CowOptions options;
+ options.op_count_max = 100000;
+ options.compression = params.compression;
+ options.num_compress_threads = params.num_threads;
+ options.batch_write = true;
+ options.compression_factor = params.block_size;
+ options.cluster_ops = params.cluster_ops;
+
+ CowWriterV3 writer(options, GetCowFd());
+
+ ASSERT_TRUE(writer.Initialize());
+
+ std::string xor_data = "This is test data-1. Testing xor";
+ xor_data.resize(options.block_size, '\0');
+ ASSERT_TRUE(writer.AddXorBlocks(50, xor_data.data(), xor_data.size(), 24, 10));
+
+ // Large number of blocks
+ std::string data = "This is test data-2. Testing replace ops";
+ data.resize(options.block_size * 2048, '\0');
+ ASSERT_TRUE(writer.AddRawBlocks(100, data.data(), data.size()));
+
+ std::string data2 = "This is test data-3. Testing replace ops";
+ data2.resize(options.block_size * 259, '\0');
+ ASSERT_TRUE(writer.AddRawBlocks(6000, data2.data(), data2.size()));
+
+ // Test data size is smaller than block size
+
+ // 4k block
+ std::string data3 = "This is test data-4. Testing replace ops";
+ data3.resize(options.block_size, '\0');
+ ASSERT_TRUE(writer.AddRawBlocks(9000, data3.data(), data3.size()));
+
+ // 8k block
+ std::string data4;
+ data4.resize(options.block_size * 2, '\0');
+ for (size_t i = 0; i < data4.size(); i++) {
+ data4[i] = static_cast<char>('A' + i / options.block_size);
+ }
+ ASSERT_TRUE(writer.AddRawBlocks(10000, data4.data(), data4.size()));
+
+ // 16k block
+ std::string data5;
+ data.resize(options.block_size * 4, '\0');
+ for (int i = 0; i < data5.size(); i++) {
+ data5[i] = static_cast<char>('C' + i / options.block_size);
+ }
+ ASSERT_TRUE(writer.AddRawBlocks(11000, data5.data(), data5.size()));
+
+ // 64k Random buffer which cannot be compressed
+ unique_fd rnd_fd(open("/dev/random", O_RDONLY));
+ ASSERT_GE(rnd_fd, 0);
+ std::string random_buffer;
+ random_buffer.resize(65536, '\0');
+ ASSERT_EQ(android::base::ReadFullyAtOffset(rnd_fd, random_buffer.data(), 65536, 0), true);
+ ASSERT_TRUE(writer.AddRawBlocks(12000, random_buffer.data(), 65536));
+
+ ASSERT_TRUE(writer.Finalize());
+
+ ASSERT_EQ(lseek(cow_->fd, 0, SEEK_SET), 0);
+
+ CowReader reader;
+ ASSERT_TRUE(reader.Parse(cow_->fd));
+
+ auto iter = reader.GetOpIter();
+ ASSERT_NE(iter, nullptr);
+
+ while (!iter->AtEnd()) {
+ auto op = iter->Get();
+
+ if (op->type() == kCowXorOp) {
+ std::string sink(xor_data.size(), '\0');
+ ASSERT_EQ(op->new_block, 50);
+ ASSERT_EQ(op->source(), 98314); // 4096 * 24 + 10
+ ReadBlockData(reader, op, sink.data(), sink.size());
+ ASSERT_EQ(sink, xor_data);
+ }
+ if (op->type() == kCowReplaceOp) {
+ if (op->new_block == 100) {
+ data.resize(options.block_size);
+ std::string sink(data.size(), '\0');
+ ReadBlockData(reader, op, sink.data(), sink.size());
+ ASSERT_EQ(sink.size(), data.size());
+ ASSERT_EQ(sink, data);
+ }
+ if (op->new_block == 6000) {
+ data2.resize(options.block_size);
+ std::string sink(data2.size(), '\0');
+ ReadBlockData(reader, op, sink.data(), sink.size());
+ ASSERT_EQ(sink, data2);
+ }
+ if (op->new_block == 9000) {
+ std::string sink(data3.size(), '\0');
+ ReadBlockData(reader, op, sink.data(), sink.size());
+ ASSERT_EQ(sink, data3);
+ }
+ if (op->new_block == 10000) {
+ data4.resize(options.block_size);
+ std::string sink(options.block_size, '\0');
+ ReadBlockData(reader, op, sink.data(), sink.size());
+ ASSERT_EQ(sink, data4);
+ }
+ if (op->new_block == 11000) {
+ data5.resize(options.block_size);
+ std::string sink(options.block_size, '\0');
+ ReadBlockData(reader, op, sink.data(), sink.size());
+ ASSERT_EQ(sink, data5);
+ }
+ if (op->new_block == 12000) {
+ random_buffer.resize(options.block_size);
+ std::string sink(options.block_size, '\0');
+ ReadBlockData(reader, op, sink.data(), sink.size());
+ ASSERT_EQ(sink, random_buffer);
+ }
+ }
+
+ iter->Next();
+ }
+}
+
+std::vector<TestParam> GetTestConfigs() {
+ std::vector<TestParam> testParams;
+
+ std::vector<int> block_sizes = {4_KiB, 8_KiB, 16_KiB, 32_KiB, 64_KiB, 128_KiB, 256_KiB};
+ std::vector<std::string> compression_algo = {"none", "lz4", "zstd", "gz"};
+ std::vector<int> threads = {1, 2};
+ // This will also test batch size
+ std::vector<size_t> cluster_ops = {1, 256};
+
+ // This should test 112 combination
+ for (auto block : block_sizes) {
+ for (auto compression : compression_algo) {
+ for (auto thread : threads) {
+ for (auto cluster : cluster_ops) {
+ TestParam param;
+ param.block_size = block;
+ param.compression = compression;
+ param.num_threads = thread;
+ param.cluster_ops = cluster;
+ testParams.push_back(std::move(param));
+ }
+ }
+ }
+ }
+
+ return testParams;
+}
+
+INSTANTIATE_TEST_SUITE_P(CompressorsWithVariableBlocks, VariableBlockTest,
+ ::testing::ValuesIn(GetTestConfigs()));
+
} // namespace snapshot
} // namespace android
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/writer_v2.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/writer_v2.cpp
index 75cd111..d0864e0 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/writer_v2.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/writer_v2.cpp
@@ -185,7 +185,7 @@
for (int i = 0; i < num_compress_threads_; i++) {
std::unique_ptr<ICompressor> compressor =
ICompressor::Create(compression_, header_.block_size);
- auto wt = std::make_unique<CompressWorker>(std::move(compressor), header_.block_size);
+ auto wt = std::make_unique<CompressWorker>(std::move(compressor));
threads_.emplace_back(std::async(std::launch::async, &CompressWorker::RunThread, wt.get()));
compress_threads_.push_back(std::move(wt));
}
@@ -353,7 +353,7 @@
if (i == num_threads - 1) {
num_blocks_per_thread = num_blocks;
}
- worker->EnqueueCompressBlocks(iter, num_blocks_per_thread);
+ worker->EnqueueCompressBlocks(iter, header_.block_size, num_blocks_per_thread);
iter += (num_blocks_per_thread * header_.block_size);
num_blocks -= num_blocks_per_thread;
}
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/writer_v2.h b/fs_mgr/libsnapshot/libsnapshot_cow/writer_v2.h
index 05de2ad..6a37aa7 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/writer_v2.h
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/writer_v2.h
@@ -81,8 +81,8 @@
int num_compress_threads_ = 1;
std::vector<std::unique_ptr<CompressWorker>> compress_threads_;
std::vector<std::future<bool>> threads_;
- std::vector<std::basic_string<uint8_t>> compressed_buf_;
- std::vector<std::basic_string<uint8_t>>::iterator buf_iter_;
+ std::vector<std::vector<uint8_t>> compressed_buf_;
+ std::vector<std::vector<uint8_t>>::iterator buf_iter_;
std::vector<std::unique_ptr<CowOperationV2>> opbuffer_vec_;
std::vector<std::unique_ptr<uint8_t[]>> databuffer_vec_;
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/writer_v3.cpp b/fs_mgr/libsnapshot/libsnapshot_cow/writer_v3.cpp
index 251b24e..de2e528 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/writer_v3.cpp
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/writer_v3.cpp
@@ -37,6 +37,7 @@
#include <libsnapshot/cow_compress.h>
#include <libsnapshot_cow/parser_v3.h>
#include <linux/fs.h>
+#include <storage_literals/storage_literals.h>
#include <sys/ioctl.h>
#include <unistd.h>
#include <numeric>
@@ -54,6 +55,7 @@
static_assert(sizeof(off_t) == sizeof(uint64_t));
+using namespace android::storage_literals;
using android::base::unique_fd;
// Divide |x| by |y| and round up to the nearest integer.
@@ -77,9 +79,9 @@
threads_.clear();
for (size_t i = 0; i < num_compress_threads_; i++) {
std::unique_ptr<ICompressor> compressor =
- ICompressor::Create(compression_, header_.block_size);
+ ICompressor::Create(compression_, header_.max_compression_size);
auto&& wt = compress_threads_.emplace_back(
- std::make_unique<CompressWorker>(std::move(compressor), header_.block_size));
+ std::make_unique<CompressWorker>(std::move(compressor)));
threads_.emplace_back(std::thread([wt = wt.get()]() { wt->RunThread(); }));
}
LOG(INFO) << num_compress_threads_ << " thread used for compression";
@@ -111,10 +113,14 @@
header_.op_count_max = 0;
header_.compression_algorithm = kCowCompressNone;
header_.max_compression_size = options_.compression_factor;
- return;
}
bool CowWriterV3::ParseOptions() {
+ if (!header_.max_compression_size || !IsBlockAligned(header_.max_compression_size)) {
+ LOG(ERROR) << "Invalid compression factor: " << header_.max_compression_size;
+ return false;
+ }
+
num_compress_threads_ = std::max(int(options_.num_compress_threads), 1);
auto parts = android::base::Split(options_.compression, ",");
if (parts.size() > 2) {
@@ -154,20 +160,22 @@
compression_.algorithm = *algorithm;
if (compression_.algorithm != kCowCompressNone) {
- compressor_ = ICompressor::Create(compression_, header_.block_size);
+ compressor_ = ICompressor::Create(compression_, header_.max_compression_size);
if (compressor_ == nullptr) {
LOG(ERROR) << "Failed to create compressor for " << compression_.algorithm;
return false;
}
- if (options_.cluster_ops &&
- (android::base::GetBoolProperty("ro.virtual_ab.batch_writes", false) ||
- options_.batch_write)) {
- batch_size_ = std::max<size_t>(options_.cluster_ops, 1);
- data_vec_.reserve(batch_size_);
- cached_data_.reserve(batch_size_);
- cached_ops_.reserve(batch_size_);
- }
}
+
+ if (options_.cluster_ops &&
+ (android::base::GetBoolProperty("ro.virtual_ab.batch_writes", false) ||
+ options_.batch_write)) {
+ batch_size_ = std::max<size_t>(options_.cluster_ops, 1);
+ data_vec_.reserve(batch_size_);
+ cached_data_.reserve(batch_size_);
+ cached_ops_.reserve(batch_size_);
+ }
+
if (batch_size_ > 1) {
LOG(INFO) << "Batch writes: enabled with batch size " << batch_size_;
} else {
@@ -178,6 +186,7 @@
num_compress_threads_ = options_.num_compress_threads;
}
InitWorkers();
+
return true;
}
@@ -206,6 +215,14 @@
}
}
+ // TODO: b/322279333
+ // Set compression factor to 4k during estimation.
+ // Once COW estimator is ready to support variable
+ // block size, this check has to be removed.
+ if (IsEstimating()) {
+ header_.max_compression_size = header_.block_size;
+ }
+
return true;
}
@@ -328,6 +345,49 @@
return cached_data_.size() >= batch_size_ || cached_ops_.size() >= batch_size_ * 16;
}
+bool CowWriterV3::ConstructCowOpCompressedBuffers(uint64_t new_block_start, const void* data,
+ uint64_t old_block, uint16_t offset,
+ CowOperationType type, size_t blocks_to_write) {
+ size_t compressed_bytes = 0;
+ auto&& blocks = CompressBlocks(blocks_to_write, data, type);
+ if (blocks.empty()) {
+ LOG(ERROR) << "Failed to compress blocks " << new_block_start << ", " << blocks_to_write
+ << ", actual number of blocks received from compressor " << blocks.size();
+ return false;
+ }
+ if (!CheckOpCount(blocks.size())) {
+ return false;
+ }
+ size_t blocks_written = 0;
+ for (size_t blk_index = 0; blk_index < blocks.size(); blk_index++) {
+ CowOperation& op = cached_ops_.emplace_back();
+ auto& vec = data_vec_.emplace_back();
+ CompressedBuffer buffer = std::move(blocks[blk_index]);
+ auto& compressed_data = cached_data_.emplace_back(std::move(buffer.compressed_data));
+ op.new_block = new_block_start + blocks_written;
+
+ op.set_type(type);
+ op.set_compression_bits(std::log2(buffer.compression_factor / header_.block_size));
+
+ if (type == kCowXorOp) {
+ op.set_source((old_block + blocks_written) * header_.block_size + offset);
+ } else {
+ op.set_source(next_data_pos_ + compressed_bytes);
+ }
+
+ vec = {.iov_base = compressed_data.data(), .iov_len = compressed_data.size()};
+ op.data_length = vec.iov_len;
+ compressed_bytes += op.data_length;
+ blocks_written += (buffer.compression_factor / header_.block_size);
+ }
+ if (blocks_written != blocks_to_write) {
+ LOG(ERROR) << "Total compressed blocks: " << blocks_written
+ << " Expected: " << blocks_to_write;
+ return false;
+ }
+ return true;
+}
+
bool CowWriterV3::EmitBlocks(uint64_t new_block_start, const void* data, size_t size,
uint64_t old_block, uint16_t offset, CowOperationType type) {
if (compression_.algorithm != kCowCompressNone && compressor_ == nullptr) {
@@ -337,42 +397,22 @@
}
const auto bytes = reinterpret_cast<const uint8_t*>(data);
const size_t num_blocks = (size / header_.block_size);
- if (!CheckOpCount(num_blocks)) {
- return false;
- }
for (size_t i = 0; i < num_blocks;) {
- const auto blocks_to_write =
+ const size_t blocks_to_write =
std::min<size_t>(batch_size_ - cached_data_.size(), num_blocks - i);
- size_t compressed_bytes = 0;
- auto&& blocks = CompressBlocks(blocks_to_write, bytes + header_.block_size * i);
- if (blocks.size() != blocks_to_write) {
- LOG(ERROR) << "Failed to compress blocks " << new_block_start + i << ", "
- << blocks_to_write << ", actual number of blocks received from compressor "
- << blocks.size();
+
+ if (!ConstructCowOpCompressedBuffers(new_block_start + i, bytes + header_.block_size * i,
+ old_block + i, offset, type, blocks_to_write)) {
return false;
}
- for (size_t j = 0; j < blocks_to_write; j++) {
- CowOperation& op = cached_ops_.emplace_back();
- auto& vec = data_vec_.emplace_back();
- auto& compressed_data = cached_data_.emplace_back(std::move(blocks[j]));
- op.new_block = new_block_start + i + j;
- op.set_type(type);
- if (type == kCowXorOp) {
- op.set_source((old_block + i + j) * header_.block_size + offset);
- } else {
- op.set_source(next_data_pos_ + compressed_bytes);
- }
- vec = {.iov_base = compressed_data.data(), .iov_len = compressed_data.size()};
- op.data_length = vec.iov_len;
- compressed_bytes += op.data_length;
- }
if (NeedsFlush() && !FlushCacheOps()) {
LOG(ERROR) << "EmitBlocks with compression: write failed. new block: "
<< new_block_start << " compression: " << compression_.algorithm
<< ", op type: " << type;
return false;
}
+
i += blocks_to_write;
}
@@ -471,8 +511,7 @@
}
bytes_written += op.data_length;
}
- if (!WriteOperation({cached_ops_.data(), cached_ops_.size()},
- {data_vec_.data(), data_vec_.size()})) {
+ if (!WriteOperation(cached_ops_, data_vec_)) {
LOG(ERROR) << "Failed to flush " << cached_ops_.size() << " ops to disk";
return false;
}
@@ -482,59 +521,175 @@
return true;
}
-std::vector<std::basic_string<uint8_t>> CowWriterV3::CompressBlocks(const size_t num_blocks,
- const void* data) {
- const size_t num_threads = (num_blocks == 1) ? 1 : num_compress_threads_;
- const size_t blocks_per_thread = DivRoundUp(num_blocks, num_threads);
- std::vector<std::basic_string<uint8_t>> compressed_buf;
- compressed_buf.clear();
- const uint8_t* const iter = reinterpret_cast<const uint8_t*>(data);
- if (compression_.algorithm == kCowCompressNone) {
- for (size_t i = 0; i < num_blocks; i++) {
- auto& buf = compressed_buf.emplace_back();
- buf.resize(header_.block_size);
- std::memcpy(buf.data(), iter + i * header_.block_size, header_.block_size);
- }
- return compressed_buf;
- }
- if (num_threads <= 1) {
- if (!CompressWorker::CompressBlocks(compressor_.get(), header_.block_size, data, num_blocks,
- &compressed_buf)) {
- return {};
- }
- } else {
- // Submit the blocks per thread. The retrieval of
- // compressed buffers has to be done in the same order.
- // We should not poll for completed buffers in a different order as the
- // buffers are tightly coupled with block ordering.
- for (size_t i = 0; i < num_threads; i++) {
- CompressWorker* worker = compress_threads_[i].get();
- const auto blocks_in_batch =
- std::min(num_blocks - i * blocks_per_thread, blocks_per_thread);
- worker->EnqueueCompressBlocks(iter + i * blocks_per_thread * header_.block_size,
- blocks_in_batch);
- }
-
- for (size_t i = 0; i < num_threads; i++) {
- CompressWorker* worker = compress_threads_[i].get();
- if (!worker->GetCompressedBuffers(&compressed_buf)) {
- return {};
- }
- }
- }
- for (size_t i = 0; i < num_blocks; i++) {
- auto& block = compressed_buf[i];
- if (block.size() >= header_.block_size) {
- block.resize(header_.block_size);
- std::memcpy(block.data(), iter + header_.block_size * i, header_.block_size);
- }
+size_t CowWriterV3::GetCompressionFactor(const size_t blocks_to_compress,
+ CowOperationType type) const {
+ // For XOR ops, we don't support bigger block size compression yet.
+ // For bigger block size support, snapshot-merge also has to changed. We
+ // aren't there yet; hence, just stick to 4k for now until
+ // snapshot-merge is ready for XOR operation.
+ if (type == kCowXorOp) {
+ return header_.block_size;
}
- return compressed_buf;
+ size_t compression_factor = header_.max_compression_size;
+ while (compression_factor > header_.block_size) {
+ size_t num_blocks = compression_factor / header_.block_size;
+ if (blocks_to_compress >= num_blocks) {
+ return compression_factor;
+ }
+ compression_factor >>= 1;
+ }
+ return header_.block_size;
}
-bool CowWriterV3::WriteOperation(std::basic_string_view<CowOperationV3> ops,
- std::basic_string_view<struct iovec> data) {
+std::vector<CowWriterV3::CompressedBuffer> CowWriterV3::ProcessBlocksWithNoCompression(
+ const size_t num_blocks, const void* data, CowOperationType type) {
+ size_t blocks_to_compress = num_blocks;
+ const uint8_t* iter = reinterpret_cast<const uint8_t*>(data);
+ std::vector<CompressedBuffer> compressed_vec;
+
+ while (blocks_to_compress) {
+ CompressedBuffer buffer;
+
+ const size_t compression_factor = GetCompressionFactor(blocks_to_compress, type);
+ size_t num_blocks = compression_factor / header_.block_size;
+
+ buffer.compression_factor = compression_factor;
+ buffer.compressed_data.resize(compression_factor);
+
+ // No compression. Just copy the data as-is.
+ std::memcpy(buffer.compressed_data.data(), iter, compression_factor);
+
+ compressed_vec.push_back(std::move(buffer));
+ blocks_to_compress -= num_blocks;
+ iter += compression_factor;
+ }
+ return compressed_vec;
+}
+
+std::vector<CowWriterV3::CompressedBuffer> CowWriterV3::ProcessBlocksWithCompression(
+ const size_t num_blocks, const void* data, CowOperationType type) {
+ size_t blocks_to_compress = num_blocks;
+ const uint8_t* iter = reinterpret_cast<const uint8_t*>(data);
+ std::vector<CompressedBuffer> compressed_vec;
+
+ while (blocks_to_compress) {
+ CompressedBuffer buffer;
+
+ const size_t compression_factor = GetCompressionFactor(blocks_to_compress, type);
+ size_t num_blocks = compression_factor / header_.block_size;
+
+ buffer.compression_factor = compression_factor;
+ // Compress the blocks
+ buffer.compressed_data = compressor_->Compress(iter, compression_factor);
+ if (buffer.compressed_data.empty()) {
+ PLOG(ERROR) << "Compression failed";
+ return {};
+ }
+
+ // Check if the buffer was indeed compressed
+ if (buffer.compressed_data.size() >= compression_factor) {
+ buffer.compressed_data.resize(compression_factor);
+ std::memcpy(buffer.compressed_data.data(), iter, compression_factor);
+ }
+
+ compressed_vec.push_back(std::move(buffer));
+ blocks_to_compress -= num_blocks;
+ iter += compression_factor;
+ }
+ return compressed_vec;
+}
+
+std::vector<CowWriterV3::CompressedBuffer> CowWriterV3::ProcessBlocksWithThreadedCompression(
+ const size_t num_blocks, const void* data, CowOperationType type) {
+ const size_t num_threads = num_compress_threads_;
+ const uint8_t* iter = reinterpret_cast<const uint8_t*>(data);
+
+ // We will alternate which thread to send compress work to. E.g. alternate between T1 and T2
+ // until all blocks are processed
+ std::vector<CompressedBuffer> compressed_vec;
+ int iteration = 0;
+ int blocks_to_compress = static_cast<int>(num_blocks);
+ while (blocks_to_compress) {
+ CompressedBuffer buffer;
+ CompressWorker* worker = compress_threads_[iteration % num_threads].get();
+
+ const size_t compression_factor = GetCompressionFactor(blocks_to_compress, type);
+ size_t num_blocks = compression_factor / header_.block_size;
+
+ worker->EnqueueCompressBlocks(iter, compression_factor, 1);
+ buffer.compression_factor = compression_factor;
+ compressed_vec.push_back(std::move(buffer));
+
+ iteration++;
+ iter += compression_factor;
+ blocks_to_compress -= num_blocks;
+ }
+
+ std::vector<std::vector<uint8_t>> compressed_buf;
+ std::vector<std::vector<std::vector<uint8_t>>> worker_buffers(num_threads);
+ compressed_buf.clear();
+ for (size_t i = 0; i < num_threads; i++) {
+ CompressWorker* worker = compress_threads_[i].get();
+ if (!worker->GetCompressedBuffers(&worker_buffers[i])) {
+ return {};
+ }
+ }
+ // compressed_vec | CB 1 | CB 2 | CB 3 | CB 4 | <-compressed buffers
+ // t1 t2 t1 t2 <- processed by these threads
+ // Ordering is important here. We need to retrieve the compressed data in the same order we
+ // processed it and assume that that we submit data beginning with the first thread and then
+ // round robin the consecutive data calls. We need to Fetch compressed buffers from the threads
+ // via the same ordering
+ for (size_t i = 0; i < compressed_vec.size(); i++) {
+ compressed_buf.emplace_back(worker_buffers[i % num_threads][i / num_threads]);
+ }
+
+ if (compressed_vec.size() != compressed_buf.size()) {
+ LOG(ERROR) << "Compressed buffer size: " << compressed_buf.size()
+ << " - Expected: " << compressed_vec.size();
+ return {};
+ }
+
+ iter = reinterpret_cast<const uint8_t*>(data);
+ // Walk through all the compressed buffers
+ for (size_t i = 0; i < compressed_buf.size(); i++) {
+ auto& buffer = compressed_vec[i];
+ auto& block = compressed_buf[i];
+ size_t block_size = buffer.compression_factor;
+ // Check if the blocks was indeed compressed
+ if (block.size() >= block_size) {
+ buffer.compressed_data.resize(block_size);
+ std::memcpy(buffer.compressed_data.data(), iter, block_size);
+ } else {
+ // Compressed block
+ buffer.compressed_data.resize(block.size());
+ std::memcpy(buffer.compressed_data.data(), block.data(), block.size());
+ }
+ iter += block_size;
+ }
+ return compressed_vec;
+}
+
+std::vector<CowWriterV3::CompressedBuffer> CowWriterV3::CompressBlocks(const size_t num_blocks,
+ const void* data,
+ CowOperationType type) {
+ if (compression_.algorithm == kCowCompressNone) {
+ return ProcessBlocksWithNoCompression(num_blocks, data, type);
+ }
+
+ const size_t num_threads = (num_blocks == 1) ? 1 : num_compress_threads_;
+
+ // If no threads are required, just compress the blocks inline.
+ if (num_threads <= 1) {
+ return ProcessBlocksWithCompression(num_blocks, data, type);
+ }
+
+ return ProcessBlocksWithThreadedCompression(num_blocks, data, type);
+}
+
+bool CowWriterV3::WriteOperation(std::span<const CowOperationV3> ops,
+ std::span<const struct iovec> data) {
const auto total_data_size =
std::transform_reduce(data.begin(), data.end(), 0, std::plus<size_t>{},
[](const struct iovec& a) { return a.iov_len; });
diff --git a/fs_mgr/libsnapshot/libsnapshot_cow/writer_v3.h b/fs_mgr/libsnapshot/libsnapshot_cow/writer_v3.h
index b19af60..48eb67b 100644
--- a/fs_mgr/libsnapshot/libsnapshot_cow/writer_v3.h
+++ b/fs_mgr/libsnapshot/libsnapshot_cow/writer_v3.h
@@ -15,15 +15,20 @@
#pragma once
#include <android-base/logging.h>
+#include <span>
#include <string_view>
#include <thread>
#include <vector>
+#include <libsnapshot/cow_format.h>
+#include <storage_literals/storage_literals.h>
#include "writer_base.h"
namespace android {
namespace snapshot {
+using namespace android::storage_literals;
+
class CowWriterV3 : public CowWriterBase {
public:
explicit CowWriterV3(const CowOptions& options, android::base::unique_fd&& fd);
@@ -43,20 +48,50 @@
virtual bool EmitSequenceData(size_t num_ops, const uint32_t* data) override;
private:
+ struct CompressedBuffer {
+ size_t compression_factor;
+ std::vector<uint8_t> compressed_data;
+ };
void SetupHeaders();
bool NeedsFlush() const;
bool ParseOptions();
bool OpenForWrite();
bool OpenForAppend(uint64_t label);
- bool WriteOperation(std::basic_string_view<CowOperationV3> op,
- std::basic_string_view<struct iovec> data);
+ bool WriteOperation(std::span<const CowOperationV3> op, std::span<const struct iovec> data);
bool EmitBlocks(uint64_t new_block_start, const void* data, size_t size, uint64_t old_block,
uint16_t offset, CowOperationType type);
+ bool ConstructCowOpCompressedBuffers(uint64_t new_block_start, const void* data,
+ uint64_t old_block, uint16_t offset, CowOperationType type,
+ size_t blocks_to_write);
bool CheckOpCount(size_t op_count);
private:
- std::vector<std::basic_string<uint8_t>> CompressBlocks(const size_t num_blocks,
- const void* data);
+ std::vector<CompressedBuffer> ProcessBlocksWithNoCompression(const size_t num_blocks,
+ const void* data,
+ CowOperationType type);
+ std::vector<CompressedBuffer> ProcessBlocksWithCompression(const size_t num_blocks,
+ const void* data,
+ CowOperationType type);
+ std::vector<CompressedBuffer> ProcessBlocksWithThreadedCompression(const size_t num_blocks,
+ const void* data,
+ CowOperationType type);
+ std::vector<CompressedBuffer> CompressBlocks(const size_t num_blocks, const void* data,
+ CowOperationType type);
+ size_t GetCompressionFactor(const size_t blocks_to_compress, CowOperationType type) const;
+
+ constexpr bool IsBlockAligned(const size_t size) {
+ // These are the only block size supported. Block size beyond 256k
+ // may impact random read performance post OTA boot.
+ const size_t values[] = {4_KiB, 8_KiB, 16_KiB, 32_KiB, 64_KiB, 128_KiB, 256_KiB};
+
+ auto it = std::lower_bound(std::begin(values), std::end(values), size);
+
+ if (it != std::end(values) && *it == size) {
+ return true;
+ }
+ return false;
+ }
+
bool ReadBackVerification();
bool FlushCacheOps();
void InitWorkers();
@@ -76,7 +111,7 @@
int num_compress_threads_ = 1;
size_t batch_size_ = 1;
std::vector<CowOperationV3> cached_ops_;
- std::vector<std::basic_string<uint8_t>> cached_data_;
+ std::vector<std::vector<uint8_t>> cached_data_;
std::vector<struct iovec> data_vec_;
std::vector<std::thread> threads_;
diff --git a/fs_mgr/libsnapshot/snapshot.cpp b/fs_mgr/libsnapshot/snapshot.cpp
index ba5fb88..e6c4de6 100644
--- a/fs_mgr/libsnapshot/snapshot.cpp
+++ b/fs_mgr/libsnapshot/snapshot.cpp
@@ -508,8 +508,6 @@
// When snapshots are on current slot, we determine the size
// of block device based on the number of COW operations. We cannot
// use base device as it will be from older image.
- size_t num_ops = 0;
- uint64_t dev_sz = 0;
unique_fd fd(open(cow_file.c_str(), O_RDONLY | O_CLOEXEC));
if (fd < 0) {
PLOG(ERROR) << "Failed to open " << cow_file;
@@ -522,13 +520,18 @@
return false;
}
+ uint64_t dev_sz = 0;
const auto& header = reader.GetHeader();
- if (header.prefix.major_version > 2) {
- LOG(ERROR) << "COW format not supported";
- return false;
+ if (header.prefix.major_version == 2) {
+ const size_t num_ops = reader.get_num_total_data_ops();
+ dev_sz = (num_ops * header.block_size);
+ } else {
+ // create_snapshot will skip in-place copy ops. Hence, fetch this
+ // information directly from v3 header.
+ const auto& v3_header = reader.header_v3();
+ dev_sz = v3_header.op_count_max * v3_header.block_size;
}
- num_ops = reader.get_num_total_data_ops();
- dev_sz = (num_ops * header.block_size);
+
base_sectors = dev_sz >> 9;
} else {
// For userspace snapshots, the size of the base device is taken as the
@@ -3242,6 +3245,8 @@
// Older OTAs don't set an explicit compression type, so default to gz.
compression_algorithm = "gz";
}
+ LOG(INFO) << "using compression algorithm: " << compression_algorithm
+ << ", max compressible block size: " << compression_factor;
}
PartitionCowCreator cow_creator{
diff --git a/fs_mgr/libsnapshot/snapshotctl.cpp b/fs_mgr/libsnapshot/snapshotctl.cpp
index 0396a55..50e9f48 100644
--- a/fs_mgr/libsnapshot/snapshotctl.cpp
+++ b/fs_mgr/libsnapshot/snapshotctl.cpp
@@ -15,6 +15,7 @@
//
#include <sysexits.h>
+#include <unistd.h>
#include <chrono>
#include <filesystem>
@@ -46,9 +47,7 @@
#include "partition_cow_creator.h"
-#ifdef SNAPSHOTCTL_USERDEBUG_OR_ENG
#include <BootControlClient.h>
-#endif
using namespace std::chrono_literals;
using namespace std::string_literals;
@@ -79,7 +78,11 @@
" revert-snapshots\n"
" Prepares devices to boot without snapshots on next boot.\n"
" This does not delete the snapshot. It only removes the indicators\n"
- " so that first stage init will not mount from snapshots.\n";
+ " so that first stage init will not mount from snapshots.\n"
+ " apply-update\n"
+ " Apply the incremental OTA update wherein the snapshots are\n"
+ " directly written to COW block device. This will bypass update-engine\n"
+ " and the device will be ready to boot from the target build.\n";
return EX_USAGE;
}
@@ -96,14 +99,22 @@
bool DeleteSnapshots();
bool CleanupSnapshot() { return sm_->PrepareDeviceToBootWithoutSnapshot(); }
bool BeginUpdate();
+ bool ApplyUpdate();
private:
std::optional<std::string> GetCowImagePath(std::string& name);
+ bool PrepareUpdate();
bool WriteSnapshotPatch(std::string cow_device, std::string patch);
+ std::string GetGroupName(const android::fs_mgr::LpMetadata& pt,
+ const std::string& partiton_name);
std::unique_ptr<SnapshotManager::LockedFile> lock_;
std::unique_ptr<SnapshotManager> sm_;
std::vector<std::future<bool>> threads_;
std::string snapshot_dir_path_;
+ std::unordered_map<std::string, chromeos_update_engine::DynamicPartitionGroup*> group_map_;
+
+ std::vector<std::string> patchfiles_;
+ chromeos_update_engine::DeltaArchiveManifest manifest_;
};
MapSnapshots::MapSnapshots(std::string path) {
@@ -115,6 +126,178 @@
snapshot_dir_path_ = path + "/";
}
+std::string MapSnapshots::GetGroupName(const android::fs_mgr::LpMetadata& pt,
+ const std::string& partition_name) {
+ std::string group_name;
+ for (const auto& partition : pt.partitions) {
+ std::string name = android::fs_mgr::GetPartitionName(partition);
+ auto suffix = android::fs_mgr::GetPartitionSlotSuffix(name);
+ std::string pname = name.substr(0, name.size() - suffix.size());
+ if (pname == partition_name) {
+ std::string group_name =
+ android::fs_mgr::GetPartitionGroupName(pt.groups[partition.group_index]);
+ return group_name.substr(0, group_name.size() - suffix.size());
+ }
+ }
+ return "";
+}
+
+bool MapSnapshots::PrepareUpdate() {
+ auto source_slot = fs_mgr_get_slot_suffix();
+ auto source_slot_number = SlotNumberForSlotSuffix(source_slot);
+ auto super_source = fs_mgr_get_super_partition_name(source_slot_number);
+
+ // Get current partition information.
+ PartitionOpener opener;
+ auto source_metadata = ReadMetadata(opener, super_source, source_slot_number);
+ if (!source_metadata) {
+ LOG(ERROR) << "Could not read source partition metadata.\n";
+ return false;
+ }
+
+ auto dap = manifest_.mutable_dynamic_partition_metadata();
+ dap->set_snapshot_enabled(true);
+ dap->set_vabc_enabled(true);
+ dap->set_vabc_compression_param("lz4");
+ dap->set_cow_version(3);
+
+ for (const auto& entry : std::filesystem::directory_iterator(snapshot_dir_path_)) {
+ if (android::base::EndsWith(entry.path().generic_string(), ".patch")) {
+ patchfiles_.push_back(android::base::Basename(entry.path().generic_string()));
+ }
+ }
+
+ for (auto& patchfile : patchfiles_) {
+ std::string parsing_file = snapshot_dir_path_ + patchfile;
+ android::base::unique_fd fd(TEMP_FAILURE_RETRY(open(parsing_file.c_str(), O_RDONLY)));
+ if (fd < 0) {
+ LOG(ERROR) << "Failed to open file: " << parsing_file;
+ return false;
+ }
+ uint64_t dev_sz = lseek(fd.get(), 0, SEEK_END);
+ if (!dev_sz) {
+ LOG(ERROR) << "Could not determine block device size: " << parsing_file;
+ return false;
+ }
+
+ const int block_sz = 4_KiB;
+ dev_sz += block_sz - 1;
+ dev_sz &= ~(block_sz - 1);
+
+ auto npos = patchfile.rfind(".patch");
+ auto partition_name = patchfile.substr(0, npos);
+
+ chromeos_update_engine::DynamicPartitionGroup* group = nullptr;
+ std::string group_name = GetGroupName(*source_metadata.get(), partition_name);
+ if (group_map_.find(group_name) != group_map_.end()) {
+ group = group_map_[group_name];
+ } else {
+ group = dap->add_groups();
+ group->set_name(group_name);
+ group_map_[group_name] = group;
+ }
+ group->add_partition_names(partition_name);
+
+ auto pu = manifest_.mutable_partitions()->Add();
+ pu->set_partition_name(partition_name);
+ pu->set_estimate_cow_size(dev_sz);
+
+ CowReader reader;
+ if (!reader.Parse(fd)) {
+ LOG(ERROR) << "COW reader parse failed";
+ return false;
+ }
+
+ uint64_t new_device_size = 0;
+ const auto& header = reader.GetHeader();
+ if (header.prefix.major_version == 2) {
+ size_t num_ops = reader.get_num_total_data_ops();
+ new_device_size = (num_ops * header.block_size);
+ } else {
+ const auto& v3_header = reader.header_v3();
+ new_device_size = v3_header.op_count_max * v3_header.block_size;
+ }
+
+ LOG(INFO) << "Partition: " << partition_name << " Group_name: " << group_name
+ << " size: " << new_device_size << " COW-size: " << dev_sz;
+ pu->mutable_new_partition_info()->set_size(new_device_size);
+ }
+ return true;
+}
+
+bool MapSnapshots::ApplyUpdate() {
+ if (!PrepareUpdate()) {
+ LOG(ERROR) << "PrepareUpdate failed";
+ return false;
+ }
+ if (!sm_->BeginUpdate()) {
+ LOG(ERROR) << "BeginUpdate failed";
+ return false;
+ }
+ if (!sm_->CreateUpdateSnapshots(manifest_)) {
+ LOG(ERROR) << "Could not apply snapshots";
+ return false;
+ }
+
+ LOG(INFO) << "CreateUpdateSnapshots success";
+ if (!sm_->MapAllSnapshots(10s)) {
+ LOG(ERROR) << "MapAllSnapshots failed";
+ return false;
+ }
+
+ LOG(INFO) << "MapAllSnapshots success";
+
+ auto& dm = android::dm::DeviceMapper::Instance();
+ auto target_slot = fs_mgr_get_other_slot_suffix();
+ for (auto& patchfile : patchfiles_) {
+ auto npos = patchfile.rfind(".patch");
+ auto partition_name = patchfile.substr(0, npos) + target_slot;
+ auto cow_device = partition_name + "-cow";
+ std::string cow_path;
+ if (!dm.GetDmDevicePathByName(cow_device, &cow_path)) {
+ LOG(ERROR) << "Failed to cow path";
+ return false;
+ }
+ threads_.emplace_back(std::async(std::launch::async, &MapSnapshots::WriteSnapshotPatch,
+ this, cow_path, patchfile));
+ }
+
+ bool ret = true;
+ for (auto& t : threads_) {
+ ret = t.get() && ret;
+ }
+ if (!ret) {
+ LOG(ERROR) << "Snapshot writes failed";
+ return false;
+ }
+ if (!sm_->UnmapAllSnapshots()) {
+ LOG(ERROR) << "UnmapAllSnapshots failed";
+ return false;
+ }
+
+ LOG(INFO) << "Pre-created snapshots successfully copied";
+ // All snapshots have been written.
+ if (!sm_->FinishedSnapshotWrites(false /* wipe */)) {
+ LOG(ERROR) << "Could not finalize snapshot writes.\n";
+ return false;
+ }
+
+ auto hal = hal::BootControlClient::WaitForService();
+ if (!hal) {
+ LOG(ERROR) << "Could not find IBootControl HAL.\n";
+ return false;
+ }
+ auto target_slot_number = SlotNumberForSlotSuffix(target_slot);
+ auto cr = hal->SetActiveBootSlot(target_slot_number);
+ if (!cr.IsOk()) {
+ LOG(ERROR) << "Could not set active boot slot: " << cr.errMsg;
+ return false;
+ }
+
+ LOG(INFO) << "ApplyUpdate success";
+ return true;
+}
+
bool MapSnapshots::BeginUpdate() {
lock_ = sm_->LockExclusive();
std::vector<std::string> snapshots;
@@ -227,11 +410,10 @@
if (file_offset >= dev_sz) {
break;
}
-
- if (fsync(cfd.get()) < 0) {
- PLOG(ERROR) << "Fsync failed at offset: " << file_offset << " size: " << to_read;
- return false;
- }
+ }
+ if (fsync(cfd.get()) < 0) {
+ PLOG(ERROR) << "Fsync failed";
+ return false;
}
return true;
}
@@ -367,6 +549,30 @@
return snapshot.DeleteSnapshots();
}
+bool ApplyUpdate(int argc, char** argv) {
+ android::base::InitLogging(argv, &android::base::KernelLogger);
+
+ // Make sure we are root.
+ if (::getuid() != 0) {
+ LOG(ERROR) << "Not running as root. Try \"adb root\" first.";
+ return EXIT_FAILURE;
+ }
+
+ if (argc < 3) {
+ std::cerr << " apply-update <directory location where snapshot patches are present>"
+ " Apply the snapshots to the COW block device\n";
+ return false;
+ }
+
+ std::string path = std::string(argv[2]);
+ MapSnapshots cow(path);
+ if (!cow.ApplyUpdate()) {
+ return false;
+ }
+ LOG(INFO) << "Apply update success. Please reboot the device";
+ return true;
+}
+
bool MapPrecreatedSnapshots(int argc, char** argv) {
android::base::InitLogging(argv, &android::base::KernelLogger);
@@ -554,6 +760,7 @@
{"test-blank-ota", TestOtaHandler},
#endif
{"unmap", UnmapCmdHandler},
+ {"apply-update", ApplyUpdate},
{"map-snapshots", MapPrecreatedSnapshots},
{"unmap-snapshots", UnMapPrecreatedSnapshots},
{"delete-snapshots", DeletePrecreatedSnapshots},
diff --git a/fs_mgr/libsnapshot/snapuserd/user-space-merge/merge_worker.cpp b/fs_mgr/libsnapshot/snapuserd/user-space-merge/merge_worker.cpp
index 1e7d0c0..bd7eaca 100644
--- a/fs_mgr/libsnapshot/snapuserd/user-space-merge/merge_worker.cpp
+++ b/fs_mgr/libsnapshot/snapuserd/user-space-merge/merge_worker.cpp
@@ -13,10 +13,11 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
-#include "merge_worker.h"
+#include <libsnapshot/cow_format.h>
#include <pthread.h>
+#include "merge_worker.h"
#include "snapuserd_core.h"
#include "utility.h"
@@ -37,6 +38,7 @@
int num_ops = *pending_ops;
int nr_consecutive = 0;
bool checkOrderedOp = (replace_zero_vec == nullptr);
+ size_t num_blocks = 1;
do {
if (!cowop_iter_->AtEnd() && num_ops) {
@@ -48,11 +50,15 @@
*source_offset = cow_op->new_block * BLOCK_SZ;
if (!checkOrderedOp) {
replace_zero_vec->push_back(cow_op);
+ if (cow_op->type() == kCowReplaceOp) {
+ // Get the number of blocks this op has compressed
+ num_blocks = (CowOpCompressionSize(cow_op, BLOCK_SZ) / BLOCK_SZ);
+ }
}
cowop_iter_->Next();
- num_ops -= 1;
- nr_consecutive = 1;
+ num_ops -= num_blocks;
+ nr_consecutive = num_blocks;
while (!cowop_iter_->AtEnd() && num_ops) {
const CowOperation* op = cowop_iter_->Get();
@@ -66,11 +72,20 @@
}
if (!checkOrderedOp) {
+ if (op->type() == kCowReplaceOp) {
+ num_blocks = (CowOpCompressionSize(op, BLOCK_SZ) / BLOCK_SZ);
+ if (num_ops < num_blocks) {
+ break;
+ }
+ } else {
+ // zero op
+ num_blocks = 1;
+ }
replace_zero_vec->push_back(op);
}
- nr_consecutive += 1;
- num_ops -= 1;
+ nr_consecutive += num_blocks;
+ num_ops -= num_blocks;
cowop_iter_->Next();
}
}
@@ -108,18 +123,24 @@
for (size_t i = 0; i < replace_zero_vec.size(); i++) {
const CowOperation* cow_op = replace_zero_vec[i];
-
- void* buffer = bufsink_.AcquireBuffer(BLOCK_SZ);
- if (!buffer) {
- SNAP_LOG(ERROR) << "AcquireBuffer failed in MergeReplaceOps";
- return false;
- }
if (cow_op->type() == kCowReplaceOp) {
- if (!reader_->ReadData(cow_op, buffer, BLOCK_SZ)) {
+ size_t buffer_size = CowOpCompressionSize(cow_op, BLOCK_SZ);
+ void* buffer = bufsink_.AcquireBuffer(buffer_size);
+ if (!buffer) {
+ SNAP_LOG(ERROR) << "AcquireBuffer failed in MergeReplaceOps";
+ return false;
+ }
+ // Read the entire compressed buffer spanning multiple blocks
+ if (!reader_->ReadData(cow_op, buffer, buffer_size)) {
SNAP_LOG(ERROR) << "Failed to read COW in merge";
return false;
}
} else {
+ void* buffer = bufsink_.AcquireBuffer(BLOCK_SZ);
+ if (!buffer) {
+ SNAP_LOG(ERROR) << "AcquireBuffer failed in MergeReplaceOps";
+ return false;
+ }
CHECK(cow_op->type() == kCowZeroOp);
memset(buffer, 0, BLOCK_SZ);
}
@@ -137,7 +158,7 @@
return false;
}
- num_ops_merged += linear_blocks;
+ num_ops_merged += replace_zero_vec.size();
if (num_ops_merged >= total_ops_merged_per_commit) {
// Flush the data
diff --git a/fs_mgr/libsnapshot/snapuserd/user-space-merge/read_worker.cpp b/fs_mgr/libsnapshot/snapuserd/user-space-merge/read_worker.cpp
index f1d4065..d40b6d1 100644
--- a/fs_mgr/libsnapshot/snapuserd/user-space-merge/read_worker.cpp
+++ b/fs_mgr/libsnapshot/snapuserd/user-space-merge/read_worker.cpp
@@ -14,10 +14,10 @@
* limitations under the License.
*/
-#include "read_worker.h"
-
+#include <libsnapshot/cow_format.h>
#include <pthread.h>
+#include "read_worker.h"
#include "snapuserd_core.h"
#include "utility.h"
@@ -48,9 +48,10 @@
// Start the replace operation. This will read the
// internal COW format and if the block is compressed,
// it will be de-compressed.
-bool ReadWorker::ProcessReplaceOp(const CowOperation* cow_op, void* buffer) {
- if (!reader_->ReadData(cow_op, buffer, BLOCK_SZ)) {
- SNAP_LOG(ERROR) << "ProcessReplaceOp failed for block " << cow_op->new_block;
+bool ReadWorker::ProcessReplaceOp(const CowOperation* cow_op, void* buffer, size_t buffer_size) {
+ if (!reader_->ReadData(cow_op, buffer, buffer_size)) {
+ SNAP_LOG(ERROR) << "ProcessReplaceOp failed for block " << cow_op->new_block
+ << " buffer_size: " << buffer_size;
return false;
}
return true;
@@ -183,7 +184,13 @@
switch (cow_op->type()) {
case kCowReplaceOp: {
- return ProcessReplaceOp(cow_op, buffer);
+ size_t buffer_size = CowOpCompressionSize(cow_op, BLOCK_SZ);
+ uint8_t chunk[buffer_size];
+ if (!ProcessReplaceOp(cow_op, chunk, buffer_size)) {
+ return false;
+ }
+ std::memcpy(buffer, chunk, BLOCK_SZ);
+ return true;
}
case kCowZeroOp: {
@@ -209,6 +216,13 @@
return false;
}
+ const size_t compression_factor = reader_->GetMaxCompressionSize();
+ if (!compression_factor) {
+ SNAP_LOG(ERROR) << "Compression factor is set to 0 which is invalid.";
+ return false;
+ }
+ decompressed_buffer_ = std::make_unique<uint8_t[]>(compression_factor);
+
backing_store_fd_.reset(open(backing_store_device_.c_str(), O_RDONLY));
if (backing_store_fd_ < 0) {
SNAP_PLOG(ERROR) << "Open Failed: " << backing_store_device_;
@@ -276,6 +290,20 @@
return true;
}
+bool ReadWorker::GetCowOpBlockOffset(const CowOperation* cow_op, uint64_t io_block,
+ off_t* block_offset) {
+ // If this is a replace op, get the block offset of this I/O
+ // block. Multi-block compression is supported only for
+ // Replace ops.
+ //
+ // Note: This can be extended when we support COPY and XOR ops down the
+ // line as the blocks are mostly contiguous.
+ if (cow_op && cow_op->type() == kCowReplaceOp) {
+ return GetBlockOffset(cow_op, io_block, BLOCK_SZ, block_offset);
+ }
+ return false;
+}
+
bool ReadWorker::ReadAlignedSector(sector_t sector, size_t sz) {
size_t remaining_size = sz;
std::vector<std::pair<sector_t, const CowOperation*>>& chunk_vec = snapuserd_->GetChunkVec();
@@ -286,7 +314,7 @@
size_t read_size = std::min(PAYLOAD_BUFFER_SZ, remaining_size);
size_t total_bytes_read = 0;
-
+ const CowOperation* prev_op = nullptr;
while (read_size) {
// We need to check every 4k block to verify if it is
// present in the mapping.
@@ -294,7 +322,7 @@
auto it = std::lower_bound(chunk_vec.begin(), chunk_vec.end(),
std::make_pair(sector, nullptr), SnapshotHandler::compare);
- bool not_found = (it == chunk_vec.end() || it->first != sector);
+ const bool sector_not_found = (it == chunk_vec.end() || it->first != sector);
void* buffer = block_server_->GetResponseBuffer(BLOCK_SZ, size);
if (!buffer) {
@@ -302,15 +330,88 @@
return false;
}
- if (not_found) {
- // Block not found in map - which means this block was not
- // changed as per the OTA. Just route the I/O to the base
- // device.
- if (!ReadDataFromBaseDevice(sector, buffer, size)) {
- SNAP_LOG(ERROR) << "ReadDataFromBaseDevice failed";
- return false;
+ if (sector_not_found) {
+ // Find the 4k block
+ uint64_t io_block = SectorToChunk(sector);
+ // Get the previous iterator. Since the vector is sorted, the
+ // lookup of this sector can fall in a range of blocks if
+ // CowOperation has compressed multiple blocks.
+ if (it != chunk_vec.begin()) {
+ std::advance(it, -1);
}
+ bool is_mapping_present = true;
+
+ // Vector itself is empty. This can happen if the block was not
+ // changed per the OTA or if the merge was already complete but
+ // snapshot table was not yet collapsed.
+ if (it == chunk_vec.end()) {
+ is_mapping_present = false;
+ }
+
+ const CowOperation* cow_op = nullptr;
+ // Relative offset within the compressed multiple blocks
+ off_t block_offset = 0;
+ if (is_mapping_present) {
+ // Get the nearest operation found in the vector
+ cow_op = it->second;
+ is_mapping_present = GetCowOpBlockOffset(cow_op, io_block, &block_offset);
+ }
+
+ // Thus, we have a case wherein sector was not found in the sorted
+ // vector; however, we indeed have a mapping of this sector
+ // embedded in one of the CowOperation which spans multiple
+ // block size.
+ if (is_mapping_present) {
+ // block_offset = 0 would mean that the CowOperation should
+ // already be in the sorted vector. Hence, lookup should
+ // have already found it. If not, this is a bug.
+ if (block_offset == 0) {
+ SNAP_LOG(ERROR)
+ << "GetBlockOffset returned offset 0 for io_block: " << io_block;
+ return false;
+ }
+
+ // Get the CowOperation actual compression size
+ size_t compression_size = CowOpCompressionSize(cow_op, BLOCK_SZ);
+ // Offset cannot be greater than the compression size
+ if (block_offset > compression_size) {
+ SNAP_LOG(ERROR) << "Invalid I/O block found. io_block: " << io_block
+ << " CowOperation-new-block: " << cow_op->new_block
+ << " compression-size: " << compression_size;
+ return false;
+ }
+
+ // Cached copy of the previous iteration. Just retrieve the
+ // data
+ if (prev_op && prev_op->new_block == cow_op->new_block) {
+ std::memcpy(buffer, (char*)decompressed_buffer_.get() + block_offset, size);
+ } else {
+ // Get the data from the disk based on the compression
+ // size
+ if (!ProcessReplaceOp(cow_op, decompressed_buffer_.get(),
+ compression_size)) {
+ return false;
+ }
+ // Copy the data from the decompressed buffer relative
+ // to the i/o block offset.
+ std::memcpy(buffer, (char*)decompressed_buffer_.get() + block_offset, size);
+ // Cache this CowOperation pointer for successive I/O
+ // operation. Since the request is sequential and the
+ // block is already decompressed, subsequest I/O blocks
+ // can fetch the data directly from this decompressed
+ // buffer.
+ prev_op = cow_op;
+ }
+ } else {
+ // Block not found in map - which means this block was not
+ // changed as per the OTA. Just route the I/O to the base
+ // device.
+ if (!ReadDataFromBaseDevice(sector, buffer, size)) {
+ SNAP_LOG(ERROR) << "ReadDataFromBaseDevice failed";
+ return false;
+ }
+ }
ret = size;
} else {
// We found the sector in mapping. Check the type of COW OP and
@@ -341,12 +442,50 @@
return true;
}
+bool ReadWorker::IsMappingPresent(const CowOperation* cow_op, loff_t requested_offset,
+ loff_t cow_op_offset) {
+ const bool replace_op = (cow_op->type() == kCowReplaceOp);
+ if (replace_op) {
+ size_t max_compressed_size = CowOpCompressionSize(cow_op, BLOCK_SZ);
+ if ((requested_offset >= cow_op_offset) &&
+ (requested_offset < (cow_op_offset + max_compressed_size))) {
+ return true;
+ }
+ }
+ return false;
+}
+
int ReadWorker::ReadUnalignedSector(
sector_t sector, size_t size,
std::vector<std::pair<sector_t, const CowOperation*>>::iterator& it) {
SNAP_LOG(DEBUG) << "ReadUnalignedSector: sector " << sector << " size: " << size
<< " Aligned sector: " << it->first;
+ loff_t requested_offset = sector << SECTOR_SHIFT;
+ loff_t final_offset = (it->first) << SECTOR_SHIFT;
+
+ const CowOperation* cow_op = it->second;
+ if (IsMappingPresent(cow_op, requested_offset, final_offset)) {
+ size_t buffer_size = CowOpCompressionSize(cow_op, BLOCK_SZ);
+ uint8_t chunk[buffer_size];
+ // Read the entire decompressed buffer based on the block-size
+ if (!ProcessReplaceOp(cow_op, chunk, buffer_size)) {
+ return -1;
+ }
+ size_t skip_offset = (requested_offset - final_offset);
+ size_t write_sz = std::min(size, buffer_size - skip_offset);
+
+ auto buffer =
+ reinterpret_cast<uint8_t*>(block_server_->GetResponseBuffer(BLOCK_SZ, write_sz));
+ if (!buffer) {
+ SNAP_LOG(ERROR) << "ReadUnalignedSector failed to allocate buffer";
+ return -1;
+ }
+
+ std::memcpy(buffer, (char*)chunk + skip_offset, write_sz);
+ return write_sz;
+ }
+
int num_sectors_skip = sector - it->first;
size_t skip_size = num_sectors_skip << SECTOR_SHIFT;
size_t write_size = std::min(size, BLOCK_SZ - skip_size);
@@ -445,8 +584,11 @@
size_t remaining_size = size;
int ret = 0;
+
+ const CowOperation* cow_op = it->second;
if (!merge_complete && (requested_offset >= final_offset) &&
- (requested_offset - final_offset) < BLOCK_SZ) {
+ (((requested_offset - final_offset) < BLOCK_SZ) ||
+ IsMappingPresent(cow_op, requested_offset, final_offset))) {
// Read the partial un-aligned data
ret = ReadUnalignedSector(sector, remaining_size, it);
if (ret < 0) {
diff --git a/fs_mgr/libsnapshot/snapuserd/user-space-merge/read_worker.h b/fs_mgr/libsnapshot/snapuserd/user-space-merge/read_worker.h
index 1aff50c..04b2736 100644
--- a/fs_mgr/libsnapshot/snapuserd/user-space-merge/read_worker.h
+++ b/fs_mgr/libsnapshot/snapuserd/user-space-merge/read_worker.h
@@ -44,9 +44,12 @@
bool ProcessXorOp(const CowOperation* cow_op, void* buffer);
bool ProcessOrderedOp(const CowOperation* cow_op, void* buffer);
bool ProcessCopyOp(const CowOperation* cow_op, void* buffer);
- bool ProcessReplaceOp(const CowOperation* cow_op, void* buffer);
+ bool ProcessReplaceOp(const CowOperation* cow_op, void* buffer, size_t buffer_size);
bool ProcessZeroOp(void* buffer);
+ bool IsMappingPresent(const CowOperation* cow_op, loff_t requested_offset,
+ loff_t cow_op_offset);
+ bool GetCowOpBlockOffset(const CowOperation* cow_op, uint64_t io_block, off_t* block_offset);
bool ReadAlignedSector(sector_t sector, size_t sz);
bool ReadUnalignedSector(sector_t sector, size_t size);
int ReadUnalignedSector(sector_t sector, size_t size,
@@ -56,6 +59,7 @@
constexpr bool IsBlockAligned(size_t size) { return ((size & (BLOCK_SZ - 1)) == 0); }
constexpr sector_t ChunkToSector(chunk_t chunk) { return chunk << CHUNK_SHIFT; }
+ constexpr chunk_t SectorToChunk(sector_t sector) { return sector >> CHUNK_SHIFT; }
std::string backing_store_device_;
unique_fd backing_store_fd_;
@@ -65,8 +69,9 @@
std::shared_ptr<IBlockServerOpener> block_server_opener_;
std::unique_ptr<IBlockServer> block_server_;
- std::basic_string<uint8_t> xor_buffer_;
+ std::vector<uint8_t> xor_buffer_;
std::unique_ptr<void, decltype(&::free)> aligned_buffer_;
+ std::unique_ptr<uint8_t[]> decompressed_buffer_;
};
} // namespace snapshot
diff --git a/fs_mgr/libsnapshot/snapuserd/user-space-merge/snapuserd_test.cpp b/fs_mgr/libsnapshot/snapuserd/user-space-merge/snapuserd_test.cpp
index 8ddb0f4..76b44b4 100644
--- a/fs_mgr/libsnapshot/snapuserd/user-space-merge/snapuserd_test.cpp
+++ b/fs_mgr/libsnapshot/snapuserd/user-space-merge/snapuserd_test.cpp
@@ -64,6 +64,9 @@
struct TestParam {
bool io_uring;
bool o_direct;
+ std::string compression;
+ int block_size;
+ int num_threads;
};
class SnapuserdTestBase : public ::testing::TestWithParam<TestParam> {
@@ -74,6 +77,7 @@
void CreateCowDevice();
void SetDeviceControlName();
std::unique_ptr<ICowWriter> CreateCowDeviceInternal();
+ std::unique_ptr<ICowWriter> CreateV3Cow();
std::unique_ptr<ITestHarness> harness_;
size_t size_ = 10_MiB;
@@ -133,6 +137,24 @@
return CreateCowWriter(kDefaultCowVersion, options, std::move(fd));
}
+std::unique_ptr<ICowWriter> SnapuserdTestBase::CreateV3Cow() {
+ const TestParam params = GetParam();
+
+ CowOptions options;
+ options.op_count_max = 100000;
+ options.compression = params.compression;
+ options.num_compress_threads = params.num_threads;
+ options.batch_write = true;
+ options.compression_factor = params.block_size;
+
+ cow_system_ = std::make_unique<TemporaryFile>();
+
+ unique_fd fd(cow_system_->fd);
+ cow_system_->fd = -1;
+
+ return CreateCowWriter(3, options, std::move(fd));
+}
+
void SnapuserdTestBase::CreateCowDevice() {
unique_fd rnd_fd;
loff_t offset = 0;
@@ -236,6 +258,7 @@
void SetupOrderedOpsInverted();
void SetupCopyOverlap_1();
void SetupCopyOverlap_2();
+ void SetupDeviceForPassthrough();
bool Merge();
void ValidateMerge();
void ReadSnapshotDeviceAndValidate();
@@ -258,6 +281,9 @@
void SimulateDaemonRestart();
+ void CreateCowDeviceWithNoBlockChanges();
+ void ValidateDeviceWithNoBlockChanges();
+
void CreateCowDeviceOrderedOps();
void CreateCowDeviceOrderedOpsInverted();
void CreateCowDeviceWithCopyOverlap_1();
@@ -307,6 +333,12 @@
ASSERT_NO_FATAL_FAILURE(SetupDaemon());
}
+void SnapuserdTest::SetupDeviceForPassthrough() {
+ ASSERT_NO_FATAL_FAILURE(CreateBaseDevice());
+ ASSERT_NO_FATAL_FAILURE(CreateCowDeviceWithNoBlockChanges());
+ ASSERT_NO_FATAL_FAILURE(SetupDaemon());
+}
+
void SnapuserdTest::SetupOrderedOpsInverted() {
ASSERT_NO_FATAL_FAILURE(CreateBaseDevice());
ASSERT_NO_FATAL_FAILURE(CreateCowDeviceOrderedOpsInverted());
@@ -480,6 +512,47 @@
}
}
+void SnapuserdTest::CreateCowDeviceWithNoBlockChanges() {
+ auto writer = CreateCowDeviceInternal();
+ ASSERT_NE(writer, nullptr);
+
+ std::unique_ptr<uint8_t[]> buffer = std::make_unique<uint8_t[]>(BLOCK_SZ);
+ std::memset(buffer.get(), 'A', BLOCK_SZ);
+
+ // This test focusses on not changing all the blocks thereby validating
+ // the pass-through I/O
+
+ // Replace the first block
+ ASSERT_TRUE(writer->AddRawBlocks(1, buffer.get(), BLOCK_SZ));
+
+ // Set zero block of Block 3
+ ASSERT_TRUE(writer->AddZeroBlocks(3, 1));
+
+ ASSERT_TRUE(writer->Finalize());
+ orig_buffer_ = std::make_unique<uint8_t[]>(total_base_size_);
+
+ // Read the entire base device
+ ASSERT_EQ(android::base::ReadFullyAtOffset(base_fd_, orig_buffer_.get(), total_base_size_, 0),
+ true);
+
+ off_t offset = BLOCK_SZ;
+ std::memcpy(orig_buffer_.get() + offset, buffer.get(), BLOCK_SZ);
+ offset = 3 * BLOCK_SZ;
+ std::memset(orig_buffer_.get() + offset, 0, BLOCK_SZ);
+}
+
+void SnapuserdTest::ValidateDeviceWithNoBlockChanges() {
+ unique_fd fd(open(dmuser_dev_->GetPath().c_str(), O_RDONLY));
+ ASSERT_GE(fd, 0);
+ std::unique_ptr<uint8_t[]> snapshot_buffer = std::make_unique<uint8_t[]>(size_);
+ std::memset(snapshot_buffer.get(), 'B', size_);
+
+ // All the I/O request should be a pass through to base device except for
+ // Block 1 and Block 3.
+ ASSERT_EQ(ReadFullyAtOffset(fd, snapshot_buffer.get(), size_, 0), true);
+ ASSERT_EQ(memcmp(snapshot_buffer.get(), orig_buffer_.get(), size_), 0);
+}
+
void SnapuserdTest::CreateCowDeviceWithCopyOverlap_1() {
auto writer = CreateCowDeviceInternal();
ASSERT_NE(writer, nullptr);
@@ -781,6 +854,20 @@
ASSERT_TRUE(Merge());
}
+TEST_P(SnapuserdTest, Snapshot_Passthrough) {
+ if (!harness_->HasUserDevice()) {
+ GTEST_SKIP() << "Skipping snapshot read; not supported";
+ }
+ ASSERT_NO_FATAL_FAILURE(SetupDeviceForPassthrough());
+ // I/O before merge
+ ASSERT_NO_FATAL_FAILURE(ValidateDeviceWithNoBlockChanges());
+ ASSERT_TRUE(Merge());
+ ValidateMerge();
+ // I/O after merge - daemon should read directly
+ // from base device
+ ASSERT_NO_FATAL_FAILURE(ValidateDeviceWithNoBlockChanges());
+}
+
TEST_P(SnapuserdTest, Snapshot_IO_TEST) {
if (!harness_->HasUserDevice()) {
GTEST_SKIP() << "Skipping snapshot read; not supported";
@@ -853,7 +940,7 @@
GTEST_SKIP() << "Skipping snapshot read; not supported";
}
ASSERT_NO_FATAL_FAILURE(SetupCopyOverlap_2());
- ASSERT_NO_FATAL_FAILURE(MergeInterruptAndValidate(2));
+ ASSERT_NO_FATAL_FAILURE(MergeInterruptFixed(300));
ValidateMerge();
}
@@ -881,11 +968,243 @@
ValidateMerge();
}
+class SnapuserdVariableBlockSizeTest : public SnapuserdTest {
+ public:
+ void SetupCowV3ForVariableBlockSize();
+ void ReadSnapshotWithVariableBlockSize();
+
+ protected:
+ void SetUp() override;
+ void TearDown() override;
+
+ void CreateV3CowDeviceForVariableBlockSize();
+};
+
+void SnapuserdVariableBlockSizeTest::SetupCowV3ForVariableBlockSize() {
+ ASSERT_NO_FATAL_FAILURE(CreateBaseDevice());
+ ASSERT_NO_FATAL_FAILURE(CreateV3CowDeviceForVariableBlockSize());
+ ASSERT_NO_FATAL_FAILURE(SetupDaemon());
+}
+
+void SnapuserdVariableBlockSizeTest::CreateV3CowDeviceForVariableBlockSize() {
+ auto writer = CreateV3Cow();
+
+ size_t total_data_to_write = size_;
+
+ size_t total_blocks_to_write = total_data_to_write / BLOCK_SZ;
+ size_t num_blocks_per_op = total_blocks_to_write / 4;
+ size_t source_block = 0;
+
+ size_t seq_len = num_blocks_per_op;
+ uint32_t sequence[seq_len];
+ size_t xor_block_start = seq_len * 3;
+ for (size_t i = 0; i < seq_len; i++) {
+ sequence[i] = xor_block_start + i;
+ }
+ ASSERT_TRUE(writer->AddSequenceData(seq_len, sequence));
+
+ size_t total_replace_blocks = num_blocks_per_op;
+ // Write some data which can be compressed
+ std::string data;
+ data.resize(total_replace_blocks * BLOCK_SZ, '\0');
+ for (size_t i = 0; i < data.size(); i++) {
+ data[i] = static_cast<char>('A' + i / BLOCK_SZ);
+ }
+ // REPLACE ops
+ ASSERT_TRUE(writer->AddRawBlocks(source_block, data.data(), data.size()));
+
+ total_blocks_to_write -= total_replace_blocks;
+ source_block = source_block + total_replace_blocks;
+
+ // ZERO ops
+ size_t total_zero_blocks = total_blocks_to_write / 3;
+ ASSERT_TRUE(writer->AddZeroBlocks(source_block, total_zero_blocks));
+
+ total_blocks_to_write -= total_zero_blocks;
+ source_block = source_block + total_zero_blocks;
+
+ // Generate some random data wherein few blocks cannot be compressed.
+ // This is to test the I/O path for those blocks which aren't compressed.
+ size_t total_random_data_blocks = total_blocks_to_write / 2;
+ unique_fd rnd_fd(open("/dev/random", O_RDONLY));
+
+ ASSERT_GE(rnd_fd, 0);
+ std::string random_buffer;
+ random_buffer.resize(total_random_data_blocks * BLOCK_SZ, '\0');
+ ASSERT_EQ(
+ android::base::ReadFullyAtOffset(rnd_fd, random_buffer.data(), random_buffer.size(), 0),
+ true);
+ // REPLACE ops
+ ASSERT_TRUE(writer->AddRawBlocks(source_block, random_buffer.data(), random_buffer.size()));
+
+ total_blocks_to_write -= total_random_data_blocks;
+ source_block = source_block + total_random_data_blocks;
+
+ // XOR ops will always be 4k blocks
+ std::string xor_buffer;
+ xor_buffer.resize(total_blocks_to_write * BLOCK_SZ, '\0');
+ for (size_t i = 0; i < xor_buffer.size(); i++) {
+ xor_buffer[i] = static_cast<char>('C' + i / BLOCK_SZ);
+ }
+ size_t xor_offset = 21;
+ std::string source_buffer;
+ source_buffer.resize(total_blocks_to_write * BLOCK_SZ, '\0');
+ ASSERT_EQ(android::base::ReadFullyAtOffset(base_fd_, source_buffer.data(), source_buffer.size(),
+ size_ + xor_offset),
+ true);
+ for (size_t i = 0; i < xor_buffer.size(); i++) {
+ xor_buffer[i] ^= source_buffer[i];
+ }
+
+ ASSERT_EQ(xor_block_start, source_block);
+
+ ASSERT_TRUE(writer->AddXorBlocks(source_block, xor_buffer.data(), xor_buffer.size(),
+ (size_ / BLOCK_SZ), xor_offset));
+ // Flush operations
+ ASSERT_TRUE(writer->Finalize());
+
+ // Construct the buffer required for validation
+ orig_buffer_ = std::make_unique<uint8_t[]>(total_base_size_);
+
+ // Read the entire base device
+ ASSERT_EQ(android::base::ReadFullyAtOffset(base_fd_, orig_buffer_.get(), total_base_size_, 0),
+ true);
+
+ // REPLACE ops which are compressed
+ std::memcpy(orig_buffer_.get(), data.data(), data.size());
+ size_t offset = data.size();
+
+ // ZERO ops
+ std::string zero_buffer(total_zero_blocks * BLOCK_SZ, 0);
+ std::memcpy((char*)orig_buffer_.get() + offset, (void*)zero_buffer.c_str(), zero_buffer.size());
+ offset += zero_buffer.size();
+
+ // REPLACE ops - Random buffers which aren't compressed
+ std::memcpy((char*)orig_buffer_.get() + offset, random_buffer.c_str(), random_buffer.size());
+ offset += random_buffer.size();
+
+ // XOR Ops which default to 4k block size compression irrespective of
+ // compression factor
+ ASSERT_EQ(android::base::ReadFullyAtOffset(base_fd_, (char*)orig_buffer_.get() + offset,
+ xor_buffer.size(), size_ + xor_offset),
+ true);
+ for (size_t i = 0; i < xor_buffer.size(); i++) {
+ orig_buffer_.get()[offset + i] = (uint8_t)(orig_buffer_.get()[offset + i] ^ xor_buffer[i]);
+ }
+}
+
+void SnapuserdVariableBlockSizeTest::ReadSnapshotWithVariableBlockSize() {
+ unique_fd fd(open(dmuser_dev_->GetPath().c_str(), O_RDONLY | O_DIRECT));
+ ASSERT_GE(fd, 0);
+
+ void* addr;
+ ssize_t page_size = getpagesize();
+ ASSERT_EQ(posix_memalign(&addr, page_size, size_), 0);
+ std::unique_ptr<void, decltype(&::free)> snapshot_buffer(addr, ::free);
+
+ const TestParam params = GetParam();
+
+ // Issue I/O request with various block sizes
+ size_t num_blocks = size_ / params.block_size;
+ off_t offset = 0;
+ for (size_t i = 0; i < num_blocks; i++) {
+ ASSERT_EQ(ReadFullyAtOffset(fd, (char*)snapshot_buffer.get() + offset, params.block_size,
+ offset),
+ true);
+ offset += params.block_size;
+ }
+ // Validate buffer
+ ASSERT_EQ(memcmp(snapshot_buffer.get(), orig_buffer_.get(), size_), 0);
+
+ // Reset the buffer
+ std::memset(snapshot_buffer.get(), 0, size_);
+
+ // Read one full chunk in a single shot and re-validate.
+ ASSERT_EQ(ReadFullyAtOffset(fd, snapshot_buffer.get(), size_, 0), true);
+ ASSERT_EQ(memcmp(snapshot_buffer.get(), orig_buffer_.get(), size_), 0);
+
+ // Reset the buffer
+ std::memset(snapshot_buffer.get(), 0, size_);
+
+ // Buffered I/O test
+ fd.reset(open(dmuser_dev_->GetPath().c_str(), O_RDONLY));
+ ASSERT_GE(fd, 0);
+
+ // Try not to cache
+ posix_fadvise(fd.get(), 0, size_, POSIX_FADV_DONTNEED);
+
+ size_t num_blocks_per_op = (size_ / BLOCK_SZ) / 4;
+ offset = num_blocks_per_op * BLOCK_SZ;
+ size_t read_size = 1019; // bytes
+ offset -= 111;
+
+ // Issue a un-aligned read which crosses the boundary between a REPLACE block and a ZERO
+ // block.
+ ASSERT_EQ(ReadFullyAtOffset(fd, snapshot_buffer.get(), read_size, offset), true);
+
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapshot_buffer.get(), (char*)orig_buffer_.get() + offset, read_size), 0);
+
+ offset = (num_blocks_per_op * 3) * BLOCK_SZ;
+ offset -= (BLOCK_SZ - 119);
+ read_size = 8111;
+
+ // Issue an un-aligned read which crosses the boundary between a REPLACE block of random
+ // un-compressed data and a XOR block
+ ASSERT_EQ(ReadFullyAtOffset(fd, snapshot_buffer.get(), read_size, offset), true);
+
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapshot_buffer.get(), (char*)orig_buffer_.get() + offset, read_size), 0);
+
+ // Reset the buffer
+ std::memset(snapshot_buffer.get(), 0, size_);
+
+ // Read just one byte at an odd offset which is a REPLACE op
+ offset = 19;
+ read_size = 1;
+ ASSERT_EQ(ReadFullyAtOffset(fd, snapshot_buffer.get(), read_size, offset), true);
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapshot_buffer.get(), (char*)orig_buffer_.get() + offset, read_size), 0);
+
+ // Reset the buffer
+ std::memset(snapshot_buffer.get(), 0, size_);
+
+ // Read a block which has no mapping to a COW operation. This read should be
+ // a pass-through to the underlying base device.
+ offset = size_ + 9342;
+ read_size = 30;
+ ASSERT_EQ(ReadFullyAtOffset(fd, snapshot_buffer.get(), read_size, offset), true);
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapshot_buffer.get(), (char*)orig_buffer_.get() + offset, read_size), 0);
+}
+
+void SnapuserdVariableBlockSizeTest::SetUp() {
+ ASSERT_NO_FATAL_FAILURE(SnapuserdTest::SetUp());
+}
+
+void SnapuserdVariableBlockSizeTest::TearDown() {
+ SnapuserdTest::TearDown();
+}
+
+TEST_P(SnapuserdVariableBlockSizeTest, Snapshot_Test_Variable_Block_Size) {
+ if (!harness_->HasUserDevice()) {
+ GTEST_SKIP() << "Skipping snapshot read; not supported";
+ }
+ ASSERT_NO_FATAL_FAILURE(SetupCowV3ForVariableBlockSize());
+ ASSERT_NO_FATAL_FAILURE(ReadSnapshotWithVariableBlockSize());
+ ASSERT_TRUE(StartMerge());
+ CheckMergeCompletion();
+ ValidateMerge();
+ ASSERT_NO_FATAL_FAILURE(ReadSnapshotWithVariableBlockSize());
+}
+
class HandlerTest : public SnapuserdTestBase {
protected:
void SetUp() override;
void TearDown() override;
+ void SetUpV2Cow();
+ void InitializeDevice();
AssertionResult ReadSectors(sector_t sector, uint64_t size, void* buffer);
TestBlockServerFactory factory_;
@@ -896,10 +1215,11 @@
std::future<bool> handler_thread_;
};
-void HandlerTest::SetUp() {
- ASSERT_NO_FATAL_FAILURE(SnapuserdTestBase::SetUp());
- ASSERT_NO_FATAL_FAILURE(CreateBaseDevice());
+void HandlerTest::SetUpV2Cow() {
ASSERT_NO_FATAL_FAILURE(CreateCowDevice());
+}
+
+void HandlerTest::InitializeDevice() {
ASSERT_NO_FATAL_FAILURE(SetDeviceControlName());
opener_ = factory_.CreateTestOpener(system_device_ctrl_name_);
@@ -921,6 +1241,13 @@
handler_thread_ = std::async(std::launch::async, &SnapshotHandler::Start, handler_.get());
}
+void HandlerTest::SetUp() {
+ ASSERT_NO_FATAL_FAILURE(SnapuserdTestBase::SetUp());
+ ASSERT_NO_FATAL_FAILURE(CreateBaseDevice());
+ ASSERT_NO_FATAL_FAILURE(SetUpV2Cow());
+ ASSERT_NO_FATAL_FAILURE(InitializeDevice());
+}
+
void HandlerTest::TearDown() {
ASSERT_TRUE(factory_.DeleteQueue(system_device_ctrl_name_));
ASSERT_TRUE(handler_thread_.get());
@@ -986,6 +1313,147 @@
ASSERT_EQ(memcmp(snapuserd_buffer.get(), orig_buffer_.get(), SECTOR_SIZE), 0);
}
+class HandlerTestV3 : public HandlerTest {
+ public:
+ void ReadSnapshotWithVariableBlockSize();
+
+ protected:
+ void SetUp() override;
+ void TearDown() override;
+ void SetUpV3Cow();
+};
+
+void HandlerTestV3::SetUp() {
+ ASSERT_NO_FATAL_FAILURE(SnapuserdTestBase::SetUp());
+ ASSERT_NO_FATAL_FAILURE(CreateBaseDevice());
+ ASSERT_NO_FATAL_FAILURE(SetUpV3Cow());
+ ASSERT_NO_FATAL_FAILURE(InitializeDevice());
+}
+
+void HandlerTestV3::TearDown() {
+ ASSERT_NO_FATAL_FAILURE(HandlerTest::TearDown());
+}
+
+void HandlerTestV3::SetUpV3Cow() {
+ auto writer = CreateV3Cow();
+
+ size_t total_data_to_write = size_;
+
+ size_t total_blocks_to_write = total_data_to_write / BLOCK_SZ;
+ size_t num_blocks_per_op = total_blocks_to_write / 4;
+ size_t source_block = 0;
+
+ size_t total_replace_blocks = num_blocks_per_op;
+ // Write some data which can be compressed
+ std::string data;
+ data.resize(total_replace_blocks * BLOCK_SZ, '\0');
+ for (size_t i = 0; i < data.size(); i++) {
+ data[i] = static_cast<char>('A' + i / BLOCK_SZ);
+ }
+ // REPLACE ops
+ ASSERT_TRUE(writer->AddRawBlocks(source_block, data.data(), data.size()));
+
+ total_blocks_to_write -= total_replace_blocks;
+ source_block = source_block + total_replace_blocks;
+
+ // ZERO ops
+ size_t total_zero_blocks = total_blocks_to_write / 3;
+ ASSERT_TRUE(writer->AddZeroBlocks(source_block, total_zero_blocks));
+
+ total_blocks_to_write -= total_zero_blocks;
+ source_block = source_block + total_zero_blocks;
+
+ // Generate some random data wherein few blocks cannot be compressed.
+ // This is to test the I/O path for those blocks which aren't compressed.
+ size_t total_random_data_blocks = total_blocks_to_write;
+ unique_fd rnd_fd(open("/dev/random", O_RDONLY));
+
+ ASSERT_GE(rnd_fd, 0);
+ std::string random_buffer;
+ random_buffer.resize(total_random_data_blocks * BLOCK_SZ, '\0');
+ ASSERT_EQ(
+ android::base::ReadFullyAtOffset(rnd_fd, random_buffer.data(), random_buffer.size(), 0),
+ true);
+ // REPLACE ops
+ ASSERT_TRUE(writer->AddRawBlocks(source_block, random_buffer.data(), random_buffer.size()));
+ // Flush operations
+ ASSERT_TRUE(writer->Finalize());
+
+ // Construct the buffer required for validation
+ orig_buffer_ = std::make_unique<uint8_t[]>(total_base_size_);
+
+ // Read the entire base device
+ ASSERT_EQ(android::base::ReadFullyAtOffset(base_fd_, orig_buffer_.get(), total_base_size_, 0),
+ true);
+
+ // REPLACE ops which are compressed
+ std::memcpy(orig_buffer_.get(), data.data(), data.size());
+ size_t offset = data.size();
+
+ // ZERO ops
+ std::string zero_buffer(total_zero_blocks * BLOCK_SZ, 0);
+ std::memcpy((char*)orig_buffer_.get() + offset, (void*)zero_buffer.c_str(), zero_buffer.size());
+ offset += zero_buffer.size();
+
+ // REPLACE ops - Random buffers which aren't compressed
+ std::memcpy((char*)orig_buffer_.get() + offset, random_buffer.c_str(), random_buffer.size());
+}
+
+TEST_P(HandlerTestV3, Read) {
+ std::unique_ptr<uint8_t[]> snapuserd_buffer = std::make_unique<uint8_t[]>(size_);
+
+ size_t read_size = SECTOR_SIZE;
+ off_t offset = 0;
+ // Read the first sector
+ ASSERT_TRUE(ReadSectors(1, read_size, snapuserd_buffer.get()));
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapuserd_buffer.get(), orig_buffer_.get(), read_size), 0);
+
+ // Read the second block at offset 7680 (Sector 15). This will map to the
+ // first COW operation for variable block size
+ offset += (((BLOCK_SZ * 2) - SECTOR_SIZE));
+ read_size = BLOCK_SZ; // Span across two REPLACE ops
+ ASSERT_TRUE(ReadSectors(offset / SECTOR_SIZE, read_size, snapuserd_buffer.get()));
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapuserd_buffer.get(), (char*)orig_buffer_.get() + offset, read_size),
+ 0);
+
+ // Fill some other data since we are going to read zero blocks
+ std::memset(snapuserd_buffer.get(), 'Z', size_);
+
+ size_t num_blocks_per_op = (size_ / BLOCK_SZ) / 4;
+ offset = num_blocks_per_op * BLOCK_SZ;
+ // Issue read spanning between a REPLACE op and ZERO ops. The starting point
+ // is the last REPLACE op at sector 5118
+ offset -= (SECTOR_SIZE * 2);
+ // This will make sure it falls back to aligned reads after reading the
+ // first unaligned block
+ read_size = BLOCK_SZ * 6;
+ ASSERT_TRUE(ReadSectors(offset / SECTOR_SIZE, read_size, snapuserd_buffer.get()));
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapuserd_buffer.get(), (char*)orig_buffer_.get() + offset, read_size),
+ 0);
+
+ // Issue I/O request at the last block. The first chunk of (SECTOR_SIZE * 2)
+ // will be from REPLACE op which has random buffers
+ offset = (size_ - (SECTOR_SIZE * 2));
+ // Request will span beyond the COW mapping, thereby fetching data from base
+ // device.
+ read_size = BLOCK_SZ * 8;
+ ASSERT_TRUE(ReadSectors(offset / SECTOR_SIZE, read_size, snapuserd_buffer.get()));
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapuserd_buffer.get(), (char*)orig_buffer_.get() + offset, read_size),
+ 0);
+
+ // Issue I/O request which are not mapped to any COW operations
+ offset = (size_ + (SECTOR_SIZE * 3));
+ read_size = BLOCK_SZ * 3;
+ ASSERT_TRUE(ReadSectors(offset / SECTOR_SIZE, read_size, snapuserd_buffer.get()));
+ // Validate the data
+ ASSERT_EQ(std::memcmp(snapuserd_buffer.get(), (char*)orig_buffer_.get() + offset, read_size),
+ 0);
+}
+
std::vector<bool> GetIoUringConfigs() {
#if __ANDROID__
if (!android::base::GetBoolProperty("ro.virtual_ab.io_uring.enabled", false)) {
@@ -1018,6 +1486,37 @@
return testParams;
}
+std::vector<TestParam> GetVariableBlockTestConfigs() {
+ std::vector<TestParam> testParams;
+
+ std::vector<int> block_sizes = {4096, 8192, 16384, 32768, 65536, 131072};
+ std::vector<std::string> compression_algo = {"none", "lz4", "zstd", "gz"};
+ std::vector<int> threads = {1, 2};
+ std::vector<bool> uring_configs = GetIoUringConfigs();
+
+ // This should test 96 combination and validates the I/O path
+ for (auto block : block_sizes) {
+ for (auto compression : compression_algo) {
+ for (auto thread : threads) {
+ for (auto io_uring : uring_configs) {
+ TestParam param;
+ param.block_size = block;
+ param.compression = compression;
+ param.num_threads = thread;
+ param.io_uring = io_uring;
+ param.o_direct = false;
+ testParams.push_back(std::move(param));
+ }
+ }
+ }
+ }
+
+ return testParams;
+}
+
+INSTANTIATE_TEST_SUITE_P(Io, SnapuserdVariableBlockSizeTest,
+ ::testing::ValuesIn(GetVariableBlockTestConfigs()));
+INSTANTIATE_TEST_SUITE_P(Io, HandlerTestV3, ::testing::ValuesIn(GetVariableBlockTestConfigs()));
INSTANTIATE_TEST_SUITE_P(Io, SnapuserdTest, ::testing::ValuesIn(GetTestConfigs()));
INSTANTIATE_TEST_SUITE_P(Io, HandlerTest, ::testing::ValuesIn(GetTestConfigs()));
diff --git a/fs_mgr/libsnapshot/tools/cow_benchmark.cpp b/fs_mgr/libsnapshot/tools/cow_benchmark.cpp
index 4d5e346..fb463c8 100644
--- a/fs_mgr/libsnapshot/tools/cow_benchmark.cpp
+++ b/fs_mgr/libsnapshot/tools/cow_benchmark.cpp
@@ -78,7 +78,7 @@
for (size_t i = 0; i < compressors.size(); i++) {
const auto start = std::chrono::steady_clock::now();
- std::basic_string<uint8_t> compressed_data =
+ std::vector<uint8_t> compressed_data =
compressors[i]->Compress(buffer.data(), buffer.size());
const auto end = std::chrono::steady_clock::now();
const auto latency =
@@ -141,13 +141,13 @@
std::vector<std::pair<double, std::string>> ratios;
for (size_t i = 0; i < compressors.size(); i++) {
- std::vector<std::basic_string<uint8_t>> compressed_data_vec;
+ std::vector<std::vector<uint8_t>> compressed_data_vec;
int num_blocks = buffer.size() / BLOCK_SZ;
const uint8_t* iter = reinterpret_cast<const uint8_t*>(buffer.data());
const auto start = std::chrono::steady_clock::now();
while (num_blocks > 0) {
- std::basic_string<uint8_t> compressed_data = compressors[i]->Compress(iter, BLOCK_SZ);
+ std::vector<uint8_t> compressed_data = compressors[i]->Compress(iter, BLOCK_SZ);
compressed_data_vec.emplace_back(compressed_data);
num_blocks--;
iter += BLOCK_SZ;
diff --git a/fs_mgr/tests/adb-remount-test.sh b/fs_mgr/tests/adb-remount-test.sh
index 7ac7a16..526c761 100755
--- a/fs_mgr/tests/adb-remount-test.sh
+++ b/fs_mgr/tests/adb-remount-test.sh
@@ -1081,7 +1081,9 @@
LOG RUN "Testing adb disable-verity -R"
T=$(adb_date)
-adb_su disable-verity -R >&2 ||
+adb_su disable-verity -R >&2
+err=${?}
+[[ ${err} -eq 0 || ${err} -eq 255 ]] ||
die -t "${T}" "disable-verity -R failed"
sleep 2
adb_wait "${ADB_WAIT}" ||
@@ -1192,7 +1194,9 @@
LOG RUN "Testing adb remount -R"
T=$(adb_date)
-adb_su remount -R </dev/null >&2 ||
+adb_su remount -R </dev/null >&2
+err=${?}
+[[ ${err} -eq 0 || ${err} -eq 255 ]] ||
die -t "${T}" "adb remount -R failed"
sleep 2
adb_wait "${ADB_WAIT}" ||
diff --git a/fs_mgr/tools/dmctl.cpp b/fs_mgr/tools/dmctl.cpp
index 7273087..9dc8c24 100644
--- a/fs_mgr/tools/dmctl.cpp
+++ b/fs_mgr/tools/dmctl.cpp
@@ -115,6 +115,21 @@
std::string block_device = NextArg();
return std::make_unique<DmTargetAndroidVerity>(start_sector, num_sectors, keyid,
block_device);
+ } else if (target_type == "striped") {
+ if (!HasArgs(3)) {
+ std::cerr << "Expected \"striped\" <block_device0> <block_device1> <chunksize>"
+ << std::endl;
+ return nullptr;
+ }
+ std::string block_device0 = NextArg();
+ std::string block_device1 = NextArg();
+ uint64_t chunk_size;
+ if (!android::base::ParseUint(NextArg(), &chunk_size)) {
+ std::cerr << "Expected start sector, got: " << PreviousArg() << std::endl;
+ return nullptr;
+ }
+ return std::make_unique<DmTargetStripe>(start_sector, num_sectors, chunk_size,
+ block_device0, block_device1);
} else if (target_type == "bow") {
if (!HasArgs(1)) {
std::cerr << "Expected \"bow\" <block_device>" << std::endl;
diff --git a/init/Android.bp b/init/Android.bp
index 181de2e..c3abefe 100644
--- a/init/Android.bp
+++ b/init/Android.bp
@@ -88,7 +88,6 @@
init_host_sources = [
"check_builtins.cpp",
"host_import_parser.cpp",
- "host_init_verifier.cpp",
]
soong_config_module_type {
@@ -321,7 +320,6 @@
visibility: ["//packages/modules/Virtualization/microdroid"],
}
-
soong_config_module_type {
name: "init_first_stage_cc_defaults",
module_type: "cc_defaults",
@@ -599,7 +597,6 @@
},
generated_headers: [
"generated_stub_builtin_function_map",
- "generated_android_ids",
],
target: {
android: {
@@ -614,13 +611,16 @@
cc_binary {
name: "host_init_verifier",
defaults: ["init_host_defaults"],
- srcs: init_common_sources + init_host_sources,
+ srcs: ["host_init_verifier.cpp"] + init_common_sources + init_host_sources,
+ generated_headers: [
+ "generated_android_ids",
+ ],
}
cc_library_host_static {
name: "libinit_host",
defaults: ["init_host_defaults"],
- srcs: init_common_sources,
+ srcs: init_common_sources + init_host_sources,
export_include_dirs: ["."],
proto: {
export_proto_headers: true,
diff --git a/init/check_builtins.cpp b/init/check_builtins.cpp
index 461ed22..9725458 100644
--- a/init/check_builtins.cpp
+++ b/init/check_builtins.cpp
@@ -28,9 +28,9 @@
#include <android-base/parsedouble.h>
#include <android-base/parseint.h>
#include <android-base/strings.h>
+#include <property_info_parser/property_info_parser.h>
#include "builtin_arguments.h"
-#include "host_init_verifier.h"
#include "interface_utils.h"
#include "property_type.h"
#include "rlimit_parser.h"
@@ -39,6 +39,9 @@
using android::base::ParseInt;
using android::base::StartsWith;
+using android::properties::BuildTrie;
+using android::properties::PropertyInfoArea;
+using android::properties::PropertyInfoEntry;
#define ReturnIfAnyArgsEmpty() \
for (const auto& arg : args) { \
@@ -50,6 +53,26 @@
namespace android {
namespace init {
+const PropertyInfoArea* property_info_area;
+
+Result<void> InitializeHostPropertyInfoArea(const std::vector<PropertyInfoEntry>& property_infos) {
+ static std::string serialized_contexts;
+ std::string trie_error;
+ if (!BuildTrie(property_infos, "u:object_r:default_prop:s0", "string", &serialized_contexts,
+ &trie_error)) {
+ return Error() << "Unable to serialize property contexts: " << trie_error;
+ }
+
+ property_info_area = reinterpret_cast<const PropertyInfoArea*>(serialized_contexts.c_str());
+ return {};
+}
+
+static Result<void> check_stub(const BuiltinArguments& args) {
+ return {};
+}
+
+#include "generated_stub_builtin_function_map.h"
+
Result<void> check_chown(const BuiltinArguments& args) {
if (!args[1].empty()) {
auto uid = DecodeUid(args[1]);
diff --git a/init/check_builtins.h b/init/check_builtins.h
index dc1b752..9b00a7c 100644
--- a/init/check_builtins.h
+++ b/init/check_builtins.h
@@ -19,6 +19,10 @@
#include "builtin_arguments.h"
#include "result.h"
+#include <vector>
+
+#include <property_info_serializer/property_info_serializer.h>
+
namespace android {
namespace init {
@@ -43,5 +47,8 @@
Result<void> check_wait(const BuiltinArguments& args);
Result<void> check_wait_for_prop(const BuiltinArguments& args);
+Result<void> InitializeHostPropertyInfoArea(
+ const std::vector<properties::PropertyInfoEntry>& property_infos);
+
} // namespace init
} // namespace android
diff --git a/init/first_stage_console.cpp b/init/first_stage_console.cpp
index 67cac19..f6f9329 100644
--- a/init/first_stage_console.cpp
+++ b/init/first_stage_console.cpp
@@ -16,6 +16,7 @@
#include "first_stage_console.h"
+#include <spawn.h>
#include <stdio.h>
#include <sys/stat.h>
#include <sys/sysmacros.h>
@@ -65,19 +66,20 @@
return true;
}
-static void RunScript() {
- LOG(INFO) << "Attempting to run /first_stage.sh...";
- pid_t pid = fork();
- if (pid != 0) {
- int status;
- waitpid(pid, &status, 0);
- LOG(INFO) << "/first_stage.sh exited with status " << status;
- return;
- }
- const char* path = "/system/bin/sh";
- const char* args[] = {path, "/first_stage.sh", nullptr};
- int rv = execv(path, const_cast<char**>(args));
- LOG(ERROR) << "unable to execv /first_stage.sh, returned " << rv << " errno " << errno;
+static pid_t SpawnImage(const char* file) {
+ const char* argv[] = {file, NULL};
+ const char* envp[] = {NULL};
+
+ char* const* argvp = const_cast<char* const*>(argv);
+ char* const* envpp = const_cast<char* const*>(envp);
+
+ pid_t pid;
+ errno = posix_spawn(&pid, argv[0], NULL, NULL, argvp, envpp);
+ if (!errno) return pid;
+
+ PLOG(ERROR) << "Failed to spawn '" << file << "'";
+
+ return (pid_t)0;
}
namespace android {
@@ -86,24 +88,28 @@
void StartConsole(const std::string& cmdline) {
bool console = KernelConsolePresent(cmdline);
// Use a simple sigchld handler -- first_stage_console doesn't need to track or log zombies
- const struct sigaction chld_act { .sa_handler = SIG_DFL, .sa_flags = SA_NOCLDWAIT };
+ const struct sigaction chld_act {
+ .sa_flags = SA_NOCLDWAIT, .sa_handler = SIG_DFL
+ };
sigaction(SIGCHLD, &chld_act, nullptr);
pid_t pid = fork();
if (pid != 0) {
- int status;
- waitpid(pid, &status, 0);
- LOG(ERROR) << "console shell exited with status " << status;
+ wait(NULL);
+ LOG(ERROR) << "console shell exited";
return;
}
if (console) console = SetupConsole();
- RunScript();
+
+ LOG(INFO) << "Attempting to run /first_stage.sh...";
+ if (SpawnImage("/first_stage.sh")) {
+ wait(NULL);
+ LOG(INFO) << "/first_stage.sh exited";
+ }
+
if (console) {
- const char* path = "/system/bin/sh";
- const char* args[] = {path, nullptr};
- int rv = execv(path, const_cast<char**>(args));
- LOG(ERROR) << "unable to execv, returned " << rv << " errno " << errno;
+ if (SpawnImage("/system/bin/sh")) wait(NULL);
}
_exit(127);
}
diff --git a/init/host_init_verifier.cpp b/init/host_init_verifier.cpp
index 662185c..f746ab9 100644
--- a/init/host_init_verifier.cpp
+++ b/init/host_init_verifier.cpp
@@ -14,8 +14,6 @@
// limitations under the License.
//
-#include "host_init_verifier.h"
-
#include <errno.h>
#include <getopt.h>
#include <pwd.h>
@@ -36,6 +34,7 @@
#include <android-base/strings.h>
#include <generated_android_ids.h>
#include <hidl/metadata.h>
+#include <property_info_parser/property_info_parser.h>
#include <property_info_serializer/property_info_serializer.h>
#include "action.h"
@@ -57,9 +56,7 @@
using android::base::ParseInt;
using android::base::ReadFileToString;
using android::base::Split;
-using android::properties::BuildTrie;
using android::properties::ParsePropertyInfoFile;
-using android::properties::PropertyInfoArea;
using android::properties::PropertyInfoEntry;
static std::vector<std::string> passwd_files;
@@ -148,12 +145,6 @@
namespace android {
namespace init {
-static Result<void> check_stub(const BuiltinArguments& args) {
- return {};
-}
-
-#include "generated_stub_builtin_function_map.h"
-
void PrintUsage() {
fprintf(stdout, R"(usage: host_init_verifier [options]
@@ -196,8 +187,6 @@
return result;
}
-const PropertyInfoArea* property_info_area;
-
void HandlePropertyContexts(const std::string& filename,
std::vector<PropertyInfoEntry>* property_infos) {
auto file_contents = std::string();
@@ -288,16 +277,11 @@
}
SetKnownInterfaces(*interface_inheritance_hierarchy_map);
- std::string serialized_contexts;
- std::string trie_error;
- if (!BuildTrie(property_infos, "u:object_r:default_prop:s0", "string", &serialized_contexts,
- &trie_error)) {
- LOG(ERROR) << "Unable to serialize property contexts: " << trie_error;
+ if (auto result = InitializeHostPropertyInfoArea(property_infos); !result.ok()) {
+ LOG(ERROR) << result.error();
return EXIT_FAILURE;
}
- property_info_area = reinterpret_cast<const PropertyInfoArea*>(serialized_contexts.c_str());
-
if (!partition_map.empty()) {
std::vector<std::string> vendor_prefixes;
for (const auto& partition : {"vendor", "odm"}) {
diff --git a/init/host_init_verifier.h b/init/host_init_verifier.h
deleted file mode 100644
index 5d24f2a..0000000
--- a/init/host_init_verifier.h
+++ /dev/null
@@ -1,27 +0,0 @@
-/*
- * Copyright (C) 2019 The Android Open Source Project
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-#pragma once
-
-#include <property_info_parser/property_info_parser.h>
-
-namespace android {
-namespace init {
-
-extern const android::properties::PropertyInfoArea* property_info_area;
-
-} // namespace init
-} // namespace android
diff --git a/init/property_service.cpp b/init/property_service.cpp
index 30ad800..bc4ef42 100644
--- a/init/property_service.cpp
+++ b/init/property_service.cpp
@@ -1104,7 +1104,8 @@
product_first_api_level = GetIntProperty("ro.build.version.sdk", __ANDROID_API_FUTURE__);
}
- vendor_api_level = std::min(vendor_api_level_of(product_first_api_level), vendor_api_level);
+ vendor_api_level =
+ std::min(AVendorSupport_getVendorApiLevelOf(product_first_api_level), vendor_api_level);
if (vendor_api_level < 0) {
LOG(ERROR) << "Unexpected vendor api level for " << VENDOR_API_LEVEL_PROP << ". Check "
diff --git a/init/snapuserd_transition.cpp b/init/snapuserd_transition.cpp
index 3a78343..dea7af9 100644
--- a/init/snapuserd_transition.cpp
+++ b/init/snapuserd_transition.cpp
@@ -195,22 +195,20 @@
return;
}
auto start = reinterpret_cast<const void*>(map.start);
- auto len = map.end - map.start;
+ uint64_t len = android::procinfo::MappedFileSize(map);
if (!len) {
return;
}
+
if (mlock(start, len) < 0) {
- LOG(ERROR) << "mlock failed, " << start << " for " << len << " bytes.";
+ PLOG(ERROR) << "\"" << map.name << "\": mlock(" << start << ", " << len
+ << ") failed: pgoff = " << map.pgoff;
ok = false;
}
};
if (!android::procinfo::ReadProcessMaps(getpid(), callback) || !ok) {
- LOG(FATAL) << "Could not process /proc/" << getpid() << "/maps file for init, "
- << "falling back to mlockall().";
- if (mlockall(MCL_CURRENT) < 0) {
- LOG(FATAL) << "mlockall failed";
- }
+ LOG(FATAL) << "Could not process /proc/" << getpid() << "/maps file for init";
}
}
diff --git a/libcutils/ashmem_test.cpp b/libcutils/ashmem_test.cpp
index d158427..571b410 100644
--- a/libcutils/ashmem_test.cpp
+++ b/libcutils/ashmem_test.cpp
@@ -22,6 +22,8 @@
#include <sys/types.h>
#include <unistd.h>
+#include <vector>
+
#include <android-base/macros.h>
#include <android-base/unique_fd.h>
#include <cutils/ashmem.h>
@@ -61,16 +63,16 @@
EXPECT_EQ(prot, ioctl(fd, ASHMEM_GET_PROT_MASK));
}
-void FillData(uint8_t* data, size_t dataLen) {
- for (size_t i = 0; i < dataLen; i++) {
+void FillData(std::vector<uint8_t>& data) {
+ for (size_t i = 0; i < data.size(); i++) {
data[i] = i & 0xFF;
}
}
TEST(AshmemTest, BasicTest) {
- constexpr size_t size = PAGE_SIZE;
- uint8_t data[size];
- FillData(data, size);
+ const size_t size = getpagesize();
+ std::vector<uint8_t> data(size);
+ FillData(data);
unique_fd fd;
ASSERT_NO_FATAL_FAILURE(TestCreateRegion(size, fd, PROT_READ | PROT_WRITE));
@@ -78,21 +80,21 @@
void* region1 = nullptr;
ASSERT_NO_FATAL_FAILURE(TestMmap(fd, size, PROT_READ | PROT_WRITE, ®ion1));
- memcpy(region1, &data, size);
- ASSERT_EQ(0, memcmp(region1, &data, size));
+ memcpy(region1, data.data(), size);
+ ASSERT_EQ(0, memcmp(region1, data.data(), size));
EXPECT_EQ(0, munmap(region1, size));
void *region2;
ASSERT_NO_FATAL_FAILURE(TestMmap(fd, size, PROT_READ, ®ion2));
- ASSERT_EQ(0, memcmp(region2, &data, size));
+ ASSERT_EQ(0, memcmp(region2, data.data(), size));
EXPECT_EQ(0, munmap(region2, size));
}
TEST(AshmemTest, ForkTest) {
- constexpr size_t size = PAGE_SIZE;
- uint8_t data[size];
- FillData(data, size);
+ const size_t size = getpagesize();
+ std::vector<uint8_t> data(size);
+ FillData(data);
unique_fd fd;
ASSERT_NO_FATAL_FAILURE(TestCreateRegion(size, fd, PROT_READ | PROT_WRITE));
@@ -100,8 +102,8 @@
void* region1 = nullptr;
ASSERT_NO_FATAL_FAILURE(TestMmap(fd, size, PROT_READ | PROT_WRITE, ®ion1));
- memcpy(region1, &data, size);
- ASSERT_EQ(0, memcmp(region1, &data, size));
+ memcpy(region1, data.data(), size);
+ ASSERT_EQ(0, memcmp(region1, data.data(), size));
EXPECT_EQ(0, munmap(region1, size));
ASSERT_EXIT(
@@ -113,7 +115,7 @@
if (region2 == MAP_FAILED) {
_exit(1);
}
- if (memcmp(region2, &data, size) != 0) {
+ if (memcmp(region2, data.data(), size) != 0) {
_exit(2);
}
memset(region2, 0, size);
@@ -122,10 +124,10 @@
},
::testing::ExitedWithCode(0), "");
- memset(&data, 0, size);
+ memset(data.data(), 0, size);
void *region2;
ASSERT_NO_FATAL_FAILURE(TestMmap(fd, size, PROT_READ | PROT_WRITE, ®ion2));
- ASSERT_EQ(0, memcmp(region2, &data, size));
+ ASSERT_EQ(0, memcmp(region2, data.data(), size));
EXPECT_EQ(0, munmap(region2, size));
}
@@ -134,18 +136,19 @@
void* region = nullptr;
// Allocate a 4-page buffer, but leave page-sized holes on either side
- constexpr size_t size = PAGE_SIZE * 4;
- constexpr size_t dataSize = PAGE_SIZE * 2;
- constexpr size_t holeSize = PAGE_SIZE;
+ const size_t pageSize = getpagesize();
+ const size_t size = pageSize * 4;
+ const size_t dataSize = pageSize * 2;
+ const size_t holeSize = pageSize;
ASSERT_NO_FATAL_FAILURE(TestCreateRegion(size, fd, PROT_READ | PROT_WRITE));
ASSERT_NO_FATAL_FAILURE(TestMmap(fd, dataSize, PROT_READ | PROT_WRITE, ®ion, holeSize));
- uint8_t data[dataSize];
- FillData(data, dataSize);
- memcpy(region, data, dataSize);
+ std::vector<uint8_t> data(dataSize);
+ FillData(data);
+ memcpy(region, data.data(), dataSize);
- constexpr off_t dataStart = holeSize;
- constexpr off_t dataEnd = dataStart + dataSize;
+ const off_t dataStart = holeSize;
+ const off_t dataEnd = dataStart + dataSize;
// The sequence of seeks below looks something like this:
//
@@ -163,9 +166,12 @@
// Expected lseek() return value
off_t ret;
} seeks[] = {
- {99, SEEK_SET, 99}, {dataStart, SEEK_CUR, dataStart + 99},
- {0, SEEK_DATA, dataStart}, {dataStart, SEEK_HOLE, dataEnd},
- {-99, SEEK_END, size - 99}, {-dataStart, SEEK_CUR, dataEnd - 99},
+ {99, SEEK_SET, 99},
+ {dataStart, SEEK_CUR, dataStart + 99},
+ {0, SEEK_DATA, dataStart},
+ {dataStart, SEEK_HOLE, dataEnd},
+ {-99, SEEK_END, static_cast<off_t>(size) - 99},
+ {-dataStart, SEEK_CUR, dataEnd - 99},
};
for (const auto& cfg : seeks) {
errno = 0;
@@ -180,7 +186,7 @@
uint8_t buf[readSize];
ASSERT_EQ(readSize, TEMP_FAILURE_RETRY(read(fd, buf, readSize)));
- EXPECT_EQ(0, memcmp(buf, data + dataOff, readSize));
+ EXPECT_EQ(0, memcmp(buf, &data[dataOff], readSize));
}
}
@@ -189,7 +195,7 @@
TEST(AshmemTest, ProtTest) {
unique_fd fd;
- constexpr size_t size = PAGE_SIZE;
+ const size_t size = getpagesize();
void *region;
ASSERT_NO_FATAL_FAILURE(TestCreateRegion(size, fd, PROT_READ));
@@ -217,7 +223,7 @@
TEST(AshmemTest, ForkProtTest) {
unique_fd fd;
- constexpr size_t size = PAGE_SIZE;
+ const size_t size = getpagesize();
int protFlags[] = { PROT_READ, PROT_WRITE };
for (size_t i = 0; i < arraysize(protFlags); i++) {
@@ -238,9 +244,9 @@
}
TEST(AshmemTest, ForkMultiRegionTest) {
- constexpr size_t size = PAGE_SIZE;
- uint8_t data[size];
- FillData(data, size);
+ const size_t size = getpagesize();
+ std::vector<uint8_t> data(size);
+ FillData(data);
constexpr int nRegions = 16;
unique_fd fd[nRegions];
@@ -248,8 +254,8 @@
ASSERT_NO_FATAL_FAILURE(TestCreateRegion(size, fd[i], PROT_READ | PROT_WRITE));
void* region = nullptr;
ASSERT_NO_FATAL_FAILURE(TestMmap(fd[i], size, PROT_READ | PROT_WRITE, ®ion));
- memcpy(region, &data, size);
- ASSERT_EQ(0, memcmp(region, &data, size));
+ memcpy(region, data.data(), size);
+ ASSERT_EQ(0, memcmp(region, data.data(), size));
EXPECT_EQ(0, munmap(region, size));
}
@@ -262,7 +268,7 @@
if (region == MAP_FAILED) {
_exit(1);
}
- if (memcmp(region, &data, size) != 0) {
+ if (memcmp(region, data.data(), size) != 0) {
munmap(region, size);
_exit(2);
}
@@ -272,11 +278,11 @@
_exit(0);
}, ::testing::ExitedWithCode(0), "");
- memset(&data, 0, size);
+ memset(data.data(), 0, size);
for (int i = 0; i < nRegions; i++) {
void *region;
ASSERT_NO_FATAL_FAILURE(TestMmap(fd[i], size, PROT_READ | PROT_WRITE, ®ion));
- ASSERT_EQ(0, memcmp(region, &data, size));
+ ASSERT_EQ(0, memcmp(region, data.data(), size));
EXPECT_EQ(0, munmap(region, size));
}
}
diff --git a/libcutils/include/private/android_filesystem_config.h b/libcutils/include/private/android_filesystem_config.h
index 8c6e548..ea61cc2 100644
--- a/libcutils/include/private/android_filesystem_config.h
+++ b/libcutils/include/private/android_filesystem_config.h
@@ -170,6 +170,7 @@
#define AID_WAKELOCK 3010 /* Allow system wakelock read/write access */
#define AID_UHID 3011 /* Allow read/write to /dev/uhid node */
#define AID_READTRACEFS 3012 /* Allow tracefs read */
+#define AID_VIRTUALMACHINE 3013 /* Allows VMs to tune for performance*/
/* The range 5000-5999 is also reserved for vendor partition. */
#define AID_OEM_RESERVED_2_START 5000
diff --git a/libprocessgroup/TEST_MAPPING b/libprocessgroup/TEST_MAPPING
new file mode 100644
index 0000000..29a9ff0
--- /dev/null
+++ b/libprocessgroup/TEST_MAPPING
@@ -0,0 +1,7 @@
+{
+ "postsubmit": [
+ {
+ "name": "StagedRollbackTest"
+ }
+ ]
+}
diff --git a/libstats/pull_rust/Android.bp b/libstats/pull_rust/Android.bp
index 4609e6b..6902026 100644
--- a/libstats/pull_rust/Android.bp
+++ b/libstats/pull_rust/Android.bp
@@ -60,8 +60,8 @@
crate_name: "statspull_rust",
srcs: ["stats_pull.rs"],
rustlibs: [
- "liblazy_static",
"liblog_rust",
+ "libonce_cell",
"libstatslog_rust_header",
"libstatspull_bindgen",
],
diff --git a/libstats/pull_rust/stats_pull.rs b/libstats/pull_rust/stats_pull.rs
index d188b5f..b2bebcc 100644
--- a/libstats/pull_rust/stats_pull.rs
+++ b/libstats/pull_rust/stats_pull.rs
@@ -14,7 +14,7 @@
//! A Rust interface for the StatsD pull API.
-use lazy_static::lazy_static;
+use once_cell::sync::Lazy;
use statslog_rust_header::{Atoms, Stat, StatsError};
use statspull_bindgen::*;
use std::collections::HashMap;
@@ -107,9 +107,8 @@
}
}
-lazy_static! {
- static ref COOKIES: Mutex<HashMap<i32, fn() -> StatsPullResult>> = Mutex::new(HashMap::new());
-}
+static COOKIES: Lazy<Mutex<HashMap<i32, fn() -> StatsPullResult>>> =
+ Lazy::new(|| Mutex::new(HashMap::new()));
/// # Safety
///
diff --git a/libstats/socket_lazy/libstatssocket_lazy.cpp b/libstats/socket_lazy/libstatssocket_lazy.cpp
index dd93eeb..fe94ef2 100644
--- a/libstats/socket_lazy/libstatssocket_lazy.cpp
+++ b/libstats/socket_lazy/libstatssocket_lazy.cpp
@@ -45,6 +45,7 @@
k_AStatsEvent_writeBool,
k_AStatsEvent_writeByteArray,
k_AStatsEvent_writeString,
+ k_AStatsEvent_writeStringArray,
k_AStatsEvent_writeAttributionChain,
k_AStatsEvent_addBoolAnnotation,
k_AStatsEvent_addInt32Annotation,
@@ -104,6 +105,7 @@
BIND_SYMBOL(AStatsEvent_writeBool);
BIND_SYMBOL(AStatsEvent_writeByteArray);
BIND_SYMBOL(AStatsEvent_writeString);
+ BIND_SYMBOL(AStatsEvent_writeStringArray);
BIND_SYMBOL(AStatsEvent_writeAttributionChain);
BIND_SYMBOL(AStatsEvent_addBoolAnnotation);
BIND_SYMBOL(AStatsEvent_addInt32Annotation);
@@ -179,6 +181,11 @@
INVOKE_METHOD(AStatsEvent_writeString, event, value);
}
+void AStatsEvent_writeStringArray(AStatsEvent* event, const char* const* elements,
+ size_t numElements) {
+ INVOKE_METHOD(AStatsEvent_writeStringArray, event, elements, numElements);
+}
+
void AStatsEvent_writeAttributionChain(AStatsEvent* event, const uint32_t* uids,
const char* const* tags, uint8_t numNodes) {
INVOKE_METHOD(AStatsEvent_writeAttributionChain, event, uids, tags, numNodes);
@@ -198,4 +205,4 @@
void AStatsSocket_close() {
INVOKE_METHOD(AStatsSocket_close);
-}
\ No newline at end of file
+}
diff --git a/libstats/socket_lazy/tests/libstatssocket_lazy_test.cpp b/libstats/socket_lazy/tests/libstatssocket_lazy_test.cpp
index fe13598..3de6cd7 100644
--- a/libstats/socket_lazy/tests/libstatssocket_lazy_test.cpp
+++ b/libstats/socket_lazy/tests/libstatssocket_lazy_test.cpp
@@ -47,6 +47,7 @@
EXPECT_DEATH(AStatsEvent_writeBool(event, false), kLoadFailed);
EXPECT_DEATH(AStatsEvent_writeByteArray(event, NULL, 0), kLoadFailed);
EXPECT_DEATH(AStatsEvent_writeString(event, NULL), kLoadFailed);
+ EXPECT_DEATH(AStatsEvent_writeStringArray(event, NULL, 0), kLoadFailed);
EXPECT_DEATH(AStatsEvent_writeAttributionChain(event, NULL, NULL, 0), kLoadFailed);
EXPECT_DEATH(AStatsEvent_addBoolAnnotation(event, 0, false), kLoadFailed);
@@ -55,4 +56,4 @@
TEST_F(LibstatssocketLazyTest, NoLibstatssocketForStatsSocket) {
EXPECT_DEATH(AStatsSocket_close(), kLoadFailed);
-}
\ No newline at end of file
+}
diff --git a/libutils/StopWatch.cpp b/libutils/StopWatch.cpp
index c88d60f..c91cd5c 100644
--- a/libutils/StopWatch.cpp
+++ b/libutils/StopWatch.cpp
@@ -18,10 +18,6 @@
#include <utils/StopWatch.h>
-/* for PRId64 */
-#ifndef __STDC_FORMAT_MACROS
-#define __STDC_FORMAT_MACROS 1
-#endif
#include <inttypes.h>
#include <log/log.h>
diff --git a/libutils/binder/RefBase_test.cpp b/libutils/binder/RefBase_test.cpp
index d675598..65d40a2 100644
--- a/libutils/binder/RefBase_test.cpp
+++ b/libutils/binder/RefBase_test.cpp
@@ -300,8 +300,8 @@
std::atomic<int>* mDeleteCount;
};
-static sp<Bar> buffer;
-static std::atomic<bool> bufferFull(false);
+[[clang::no_destroy]] static constinit sp<Bar> buffer;
+static constinit std::atomic<bool> bufferFull(false);
// Wait until bufferFull has value val.
static inline void waitFor(bool val) {
@@ -380,8 +380,8 @@
} // Otherwise this is slow and probably pointless on a uniprocessor.
}
-static wp<Bar> wpBuffer;
-static std::atomic<bool> wpBufferFull(false);
+[[clang::no_destroy]] static constinit wp<Bar> wpBuffer;
+static constinit std::atomic<bool> wpBufferFull(false);
// Wait until wpBufferFull has value val.
static inline void wpWaitFor(bool val) {
diff --git a/libutils/binder/SharedBuffer_test.cpp b/libutils/binder/SharedBuffer_test.cpp
index 1d6317f..26702b0 100644
--- a/libutils/binder/SharedBuffer_test.cpp
+++ b/libutils/binder/SharedBuffer_test.cpp
@@ -14,8 +14,6 @@
* limitations under the License.
*/
-#define __STDC_LIMIT_MACROS
-
#include <gtest/gtest.h>
#include <memory>
diff --git a/libutils/binder/String8.cpp b/libutils/binder/String8.cpp
index 749bfcb..1de9e8b 100644
--- a/libutils/binder/String8.cpp
+++ b/libutils/binder/String8.cpp
@@ -14,15 +14,13 @@
* limitations under the License.
*/
-#define __STDC_LIMIT_MACROS
-#include <stdint.h>
-
#include <utils/String8.h>
#include <log/log.h>
#include <utils/String16.h>
#include <ctype.h>
+#include <stdint.h>
#include <limits>
#include <string>
diff --git a/libutils/binder/Vector_test.cpp b/libutils/binder/Vector_test.cpp
index 6d90eaa..312dcf6 100644
--- a/libutils/binder/Vector_test.cpp
+++ b/libutils/binder/Vector_test.cpp
@@ -16,7 +16,6 @@
#define LOG_TAG "Vector_test"
-#define __STDC_LIMIT_MACROS
#include <stdint.h>
#include <unistd.h>
diff --git a/libutils/binder/include/utils/RefBase.h b/libutils/binder/include/utils/RefBase.h
index 5e3fa7d..f03e1be 100644
--- a/libutils/binder/include/utils/RefBase.h
+++ b/libutils/binder/include/utils/RefBase.h
@@ -404,7 +404,7 @@
public:
typedef typename RefBase::weakref_type weakref_type;
- inline wp() : m_ptr(nullptr), m_refs(nullptr) { }
+ inline constexpr wp() : m_ptr(nullptr), m_refs(nullptr) { }
// if nullptr, returns nullptr
//
diff --git a/libutils/binder/include/utils/StrongPointer.h b/libutils/binder/include/utils/StrongPointer.h
index 43c00c9..fb9b8e8 100644
--- a/libutils/binder/include/utils/StrongPointer.h
+++ b/libutils/binder/include/utils/StrongPointer.h
@@ -30,7 +30,7 @@
template<typename T>
class sp {
public:
- inline sp() : m_ptr(nullptr) { }
+ inline constexpr sp() : m_ptr(nullptr) { }
// The old way of using sp<> was like this. This is bad because it relies
// on implicit conversion to sp<>, which we would like to remove (if an
diff --git a/libutils/include/utils/CallStack.h b/libutils/include/utils/CallStack.h
index fe4d4f5..0239b68 100644
--- a/libutils/include/utils/CallStack.h
+++ b/libutils/include/utils/CallStack.h
@@ -14,8 +14,7 @@
* limitations under the License.
*/
-#ifndef ANDROID_CALLSTACK_H
-#define ANDROID_CALLSTACK_H
+#pragma once
#include <memory>
@@ -27,17 +26,19 @@
#include <sys/types.h>
#if !defined(__APPLE__) && !defined(_WIN32)
-# define WEAKS_AVAILABLE 1
+# define CALLSTACK_WEAKS_AVAILABLE 1
#endif
#ifndef CALLSTACK_WEAK
-# ifdef WEAKS_AVAILABLE
+# ifdef CALLSTACK_WEAKS_AVAILABLE
# define CALLSTACK_WEAK __attribute__((weak))
-# else // !WEAKS_AVAILABLE
+# else // !CALLSTACK_WEAKS_AVAILABLE
# define CALLSTACK_WEAK
-# endif // !WEAKS_AVAILABLE
+# endif // !CALLSTACK_WEAKS_AVAILABLE
#endif // CALLSTACK_WEAK predefined
-#define ALWAYS_INLINE __attribute__((always_inline))
+#ifndef CALLSTACK_ALWAYS_INLINE
+#define CALLSTACK_ALWAYS_INLINE __attribute__((always_inline))
+#endif // CALLSTACK_ALWAYS_INLINE predefined
namespace android {
@@ -89,7 +90,7 @@
//
// DO NOT USE THESE. They will disappear.
struct StackDeleter {
-#ifdef WEAKS_AVAILABLE
+#ifdef CALLSTACK_WEAKS_AVAILABLE
void operator()(CallStack* stack) {
deleteStack(stack);
}
@@ -101,8 +102,8 @@
typedef std::unique_ptr<CallStack, StackDeleter> CallStackUPtr;
// Return current call stack if possible, nullptr otherwise.
-#ifdef WEAKS_AVAILABLE
- static CallStackUPtr ALWAYS_INLINE getCurrent(int32_t ignoreDepth = 1) {
+#ifdef CALLSTACK_WEAKS_AVAILABLE
+ static CallStackUPtr CALLSTACK_ALWAYS_INLINE getCurrent(int32_t ignoreDepth = 1) {
if (reinterpret_cast<uintptr_t>(getCurrentInternal) == 0) {
ALOGW("CallStack::getCurrentInternal not linked, returning null");
return CallStackUPtr(nullptr);
@@ -110,15 +111,16 @@
return getCurrentInternal(ignoreDepth);
}
}
-#else // !WEAKS_AVAILABLE
- static CallStackUPtr ALWAYS_INLINE getCurrent(int32_t = 1) {
+#else // !CALLSTACK_WEAKS_AVAILABLE
+ static CallStackUPtr CALLSTACK_ALWAYS_INLINE getCurrent(int32_t = 1) {
return CallStackUPtr(nullptr);
}
-#endif // !WEAKS_AVAILABLE
+#endif // !CALLSTACK_WEAKS_AVAILABLE
-#ifdef WEAKS_AVAILABLE
- static void ALWAYS_INLINE logStack(const char* logtag, CallStack* stack = getCurrent().get(),
- android_LogPriority priority = ANDROID_LOG_DEBUG) {
+#ifdef CALLSTACK_WEAKS_AVAILABLE
+ static void CALLSTACK_ALWAYS_INLINE logStack(const char* logtag,
+ CallStack* stack = getCurrent().get(),
+ android_LogPriority priority = ANDROID_LOG_DEBUG) {
if (reinterpret_cast<uintptr_t>(logStackInternal) != 0 && stack != nullptr) {
logStackInternal(logtag, stack, priority);
} else {
@@ -127,30 +129,31 @@
}
#else
- static void ALWAYS_INLINE logStack(const char* logtag, CallStack* = getCurrent().get(),
- android_LogPriority = ANDROID_LOG_DEBUG) {
+ static void CALLSTACK_ALWAYS_INLINE logStack(const char* logtag,
+ CallStack* = getCurrent().get(),
+ android_LogPriority = ANDROID_LOG_DEBUG) {
ALOG(LOG_WARN, logtag, "CallStack::logStackInternal not linked");
}
-#endif // !WEAKS_AVAILABLE
+#endif // !CALLSTACK_WEAKS_AVAILABLE
-#ifdef WEAKS_AVAILABLE
- static String8 ALWAYS_INLINE stackToString(const char* prefix = nullptr,
- const CallStack* stack = getCurrent().get()) {
+#ifdef CALLSTACK_WEAKS_AVAILABLE
+ static String8 CALLSTACK_ALWAYS_INLINE
+ stackToString(const char* prefix = nullptr, const CallStack* stack = getCurrent().get()) {
if (reinterpret_cast<uintptr_t>(stackToStringInternal) != 0 && stack != nullptr) {
return stackToStringInternal(prefix, stack);
} else {
return String8::format("%s<CallStack package not linked>", (prefix ? prefix : ""));
}
}
-#else // !WEAKS_AVAILABLE
- static String8 ALWAYS_INLINE stackToString(const char* prefix = nullptr,
- const CallStack* = getCurrent().get()) {
+#else // !CALLSTACK_WEAKS_AVAILABLE
+ static String8 CALLSTACK_ALWAYS_INLINE stackToString(const char* prefix = nullptr,
+ const CallStack* = getCurrent().get()) {
return String8::format("%s<CallStack package not linked>", (prefix ? prefix : ""));
}
-#endif // !WEAKS_AVAILABLE
+#endif // !CALLSTACK_WEAKS_AVAILABLE
private:
-#ifdef WEAKS_AVAILABLE
+#ifdef CALLSTACK_WEAKS_AVAILABLE
static CallStackUPtr CALLSTACK_WEAK getCurrentInternal(int32_t ignoreDepth);
static void CALLSTACK_WEAK logStackInternal(const char* logtag, const CallStack* stack,
android_LogPriority priority);
@@ -158,11 +161,13 @@
// The deleter is only invoked on non-null pointers. Hence it will never be
// invoked if CallStack is not linked.
static void CALLSTACK_WEAK deleteStack(CallStack* stack);
-#endif // WEAKS_AVAILABLE
+#endif // CALLSTACK_WEAKS_AVAILABLE
Vector<String8> mFrameLines;
};
} // namespace android
-#endif // ANDROID_CALLSTACK_H
+#undef CALLSTACK_WEAKS_AVAILABLE
+#undef CALLSTACK_WEAK
+#undef CALLSTACK_ALWAYS_INLINE
diff --git a/libvendorsupport/Android.bp b/libvendorsupport/Android.bp
index b4457b1..e87959e 100644
--- a/libvendorsupport/Android.bp
+++ b/libvendorsupport/Android.bp
@@ -34,3 +34,32 @@
"liblog",
],
}
+
+cc_library_headers {
+ name: "libvendorsupport_llndk_headers",
+ host_supported: true,
+ vendor_available: true,
+ recovery_available: true,
+ ramdisk_available: true,
+ vendor_ramdisk_available: true,
+ native_bridge_supported: true,
+
+ export_include_dirs: ["include_llndk"],
+ llndk: {
+ llndk_headers: true,
+ },
+
+ apex_available: [
+ "//apex_available:platform",
+ "//apex_available:anyapex",
+ ],
+ min_sdk_version: "apex_inherit",
+
+ system_shared_libs: [],
+ stl: "none",
+
+ // This header library is used for libc and must be available to any sdk
+ // versions.
+ // Setting sdk_version to the lowest version allows the dependencies.
+ sdk_version: "1",
+}
diff --git a/libvendorsupport/include/vendorsupport/api_level.h b/libvendorsupport/include/vendorsupport/api_level.h
index ba1a6b8..d365075 100644
--- a/libvendorsupport/include/vendorsupport/api_level.h
+++ b/libvendorsupport/include/vendorsupport/api_level.h
@@ -14,38 +14,34 @@
#pragma once
-#include <android/api-level.h>
+#include <sys/cdefs.h>
+
+__BEGIN_DECLS
#define __ANDROID_VENDOR_API_MAX__ 1000000
#define __INVALID_API_LEVEL -1
-#ifdef __cplusplus
-extern "C" {
-#endif
-
/**
* @brief Find corresponding vendor API level from an SDK API version.
*
* @details
* SDK API versions and vendor API levels are not compatible and not
- * convertible. However, this function can be used to compare the two versions
+ * exchangeable. However, this function can be used to compare the two versions
* to know which one is newer than the other.
*
- * @param sdk_api_level The SDK version int. This must be less than 10000.
+ * @param sdkApiLevel The SDK version int. This must be less than 10000.
* @return The corresponding vendor API level of the SDK version. -1 if the SDK
* version is invalid or 10000.
*/
-int vendor_api_level_of(int sdk_api_level);
+int AVendorSupport_getVendorApiLevelOf(int sdkApiLevel);
/**
* @brief Find corresponding SDK API version from a vendor API level.
*
- * @param vendor_api_level The vendor API level int.
+ * @param vendorApiLevel The vendor API level int.
* @return The corresponding SDK API version of the vendor API level. -1 if the
* vendor API level is invalid.
*/
-int sdk_api_level_of(int vendor_api_level);
+int AVendorSupport_getSdkApiLevelOf(int vendorApiLevel);
-#ifdef __cplusplus
-}
-#endif
+__END_DECLS
diff --git a/libvendorsupport/include_llndk/android/llndk-versioning.h b/libvendorsupport/include_llndk/android/llndk-versioning.h
new file mode 100644
index 0000000..7c408c9
--- /dev/null
+++ b/libvendorsupport/include_llndk/android/llndk-versioning.h
@@ -0,0 +1,48 @@
+// Copyright (C) 2024 The Android Open Source Project
+//
+// Licensed under the Apache License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+//
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+
+#pragma once
+
+#include <sys/cdefs.h>
+
+__BEGIN_DECLS
+
+#if defined(__ANDROID_VENDOR__)
+
+// LLNDK (https://source.android.com/docs/core/architecture/vndk/build-system#ll-ndk) is similar to
+// NDK, but uses its own versioning of YYYYMM format for vendor builds. The LLNDK symbols are
+// enabled when the vendor api level is equal to or newer than the ro.board.api_level.
+#define __INTRODUCED_IN_LLNDK(vendor_api_level) \
+ _Pragma("clang diagnostic push") _Pragma("clang diagnostic ignored \"-Wgcc-compat\"") \
+ __attribute__((enable_if( \
+ __ANDROID_VENDOR_API__ >= vendor_api_level, \
+ "available in vendor API level " #vendor_api_level " that " \
+ "is newer than the current vendor API level. Guard the API " \
+ "call with '#if (__ANDROID_VENDOR_API__ >= " #vendor_api_level ")'."))) \
+ _Pragma("clang diagnostic pop")
+
+// For the vendor libraries, __INTRODUCED_IN must be ignored because they are only for NDKs but not
+// for LLNDKs.
+#undef __INTRODUCED_IN
+#define __INTRODUCED_IN(x)
+
+#else // __ANDROID_VENDOR__
+
+// For non-vendor libraries, __INTRODUCED_IN_LLNDK must be ignored because it must not change
+// symbols of NDK or the system side of the treble boundary.
+#define __INTRODUCED_IN_LLNDK(vendor_api_level)
+
+#endif // __ANDROID_VENDOR__
+
+__END_DECLS
diff --git a/libvendorsupport/libvendorsupport.map.txt b/libvendorsupport/libvendorsupport.map.txt
index 9a23b94..d99c834 100644
--- a/libvendorsupport/libvendorsupport.map.txt
+++ b/libvendorsupport/libvendorsupport.map.txt
@@ -1,7 +1,7 @@
LIBVENDORSUPPORT {
global:
- vendor_api_level_of; # llndk systemapi
- sdk_api_level_of; # llndk systemapi
+ AVendorSupport_getVendorApiLevelOf; # llndk systemapi
+ AVendorSupport_getSdkApiLevelOf; # llndk systemapi
local:
*;
};
diff --git a/libvendorsupport/tests/version_props_test.cpp b/libvendorsupport/tests/version_props_test.cpp
index 538a2e2..ad54c88 100644
--- a/libvendorsupport/tests/version_props_test.cpp
+++ b/libvendorsupport/tests/version_props_test.cpp
@@ -21,17 +21,17 @@
namespace {
-TEST(vendorsupport, get_corresponding_vendor_api_level) {
- ASSERT_EQ(__ANDROID_API_U__, vendor_api_level_of(__ANDROID_API_U__));
- ASSERT_EQ(202404, vendor_api_level_of(__ANDROID_API_V__));
- ASSERT_EQ(__INVALID_API_LEVEL, vendor_api_level_of(__ANDROID_API_FUTURE__));
+TEST(VendorSupport, GetCorrespondingVendorApiLevel) {
+ ASSERT_EQ(__ANDROID_API_U__, AVendorSupport_getVendorApiLevelOf(__ANDROID_API_U__));
+ ASSERT_EQ(202404, AVendorSupport_getVendorApiLevelOf(__ANDROID_API_V__));
+ ASSERT_EQ(__INVALID_API_LEVEL, AVendorSupport_getVendorApiLevelOf(__ANDROID_API_FUTURE__));
}
-TEST(vendorsupport, get_corresponding_sdk_api_level) {
- ASSERT_EQ(__ANDROID_API_U__, sdk_api_level_of(__ANDROID_API_U__));
- ASSERT_EQ(__ANDROID_API_V__, sdk_api_level_of(202404));
- ASSERT_EQ(__INVALID_API_LEVEL, sdk_api_level_of(__ANDROID_VENDOR_API_MAX__));
- ASSERT_EQ(__INVALID_API_LEVEL, sdk_api_level_of(35));
+TEST(VendorSupport, GetCorrespondingSdkApiLevel) {
+ ASSERT_EQ(__ANDROID_API_U__, AVendorSupport_getSdkApiLevelOf(__ANDROID_API_U__));
+ ASSERT_EQ(__ANDROID_API_V__, AVendorSupport_getSdkApiLevelOf(202404));
+ ASSERT_EQ(__INVALID_API_LEVEL, AVendorSupport_getSdkApiLevelOf(__ANDROID_VENDOR_API_MAX__));
+ ASSERT_EQ(__INVALID_API_LEVEL, AVendorSupport_getSdkApiLevelOf(35));
}
} // namespace
\ No newline at end of file
diff --git a/libvendorsupport/version_props.c b/libvendorsupport/version_props.c
index 4d0e45e..835828c 100644
--- a/libvendorsupport/version_props.c
+++ b/libvendorsupport/version_props.c
@@ -16,26 +16,26 @@
#include <log/log.h>
-int vendor_api_level_of(int sdk_api_level) {
- if (sdk_api_level < __ANDROID_API_V__) {
- return sdk_api_level;
+int AVendorSupport_getVendorApiLevelOf(int sdkApiLevel) {
+ if (sdkApiLevel < __ANDROID_API_V__) {
+ return sdkApiLevel;
}
// In Android V, vendor API level started with version 202404.
// The calculation assumes that the SDK api level bumps once a year.
- if (sdk_api_level < __ANDROID_API_FUTURE__) {
- return 202404 + ((sdk_api_level - __ANDROID_API_V__) * 100);
+ if (sdkApiLevel < __ANDROID_API_FUTURE__) {
+ return 202404 + ((sdkApiLevel - __ANDROID_API_V__) * 100);
}
- ALOGE("The SDK version must be less than 10000: %d", sdk_api_level);
+ ALOGE("The SDK version must be less than 10000: %d", sdkApiLevel);
return __INVALID_API_LEVEL;
}
-int sdk_api_level_of(int vendor_api_level) {
- if (vendor_api_level < __ANDROID_API_V__) {
- return vendor_api_level;
+int AVendorSupport_getSdkApiLevelOf(int vendorApiLevel) {
+ if (vendorApiLevel < __ANDROID_API_V__) {
+ return vendorApiLevel;
}
- if (vendor_api_level >= 202404 && vendor_api_level < __ANDROID_VENDOR_API_MAX__) {
- return (vendor_api_level - 202404) / 100 + __ANDROID_API_V__;
+ if (vendorApiLevel >= 202404 && vendorApiLevel < __ANDROID_VENDOR_API_MAX__) {
+ return (vendorApiLevel - 202404) / 100 + __ANDROID_API_V__;
}
- ALOGE("Unexpected vendor api level: %d", vendor_api_level);
+ ALOGE("Unexpected vendor api level: %d", vendorApiLevel);
return __INVALID_API_LEVEL;
}
diff --git a/rootdir/init.rc b/rootdir/init.rc
index 4c07c83..8a2ed9f 100644
--- a/rootdir/init.rc
+++ b/rootdir/init.rc
@@ -247,6 +247,7 @@
write /dev/blkio/background/blkio.bfq.weight 10
write /dev/blkio/blkio.group_idle 0
write /dev/blkio/background/blkio.group_idle 0
+ write /dev/blkio/background/blkio.prio.class restrict-to-be
restorecon_recursive /mnt
@@ -639,6 +640,11 @@
restorecon_recursive /metadata/apex
mkdir /metadata/staged-install 0770 root system
+
+ mkdir /metadata/aconfig 0775 root system
+ mkdir /metadata/aconfig/flags 0770 root system
+ mkdir /metadata/aconfig/boot 0775 root system
+
on late-fs
# Ensure that tracefs has the correct permissions.
# This does not work correctly if it is called in post-fs.
@@ -703,7 +709,7 @@
# Start tombstoned early to be able to store tombstones.
mkdir /data/anr 0775 system system encryption=Require
- mkdir /data/tombstones 0771 system system encryption=Require
+ mkdir /data/tombstones 0775 system system encryption=Require
mkdir /data/vendor/tombstones 0771 root root
mkdir /data/vendor/tombstones/wifi 0771 wifi wifi
start tombstoned
diff --git a/trusty/keymint/src/keymint_hal_main.rs b/trusty/keymint/src/keymint_hal_main.rs
index cfa859f..eda986a 100644
--- a/trusty/keymint/src/keymint_hal_main.rs
+++ b/trusty/keymint/src/keymint_hal_main.rs
@@ -92,8 +92,8 @@
android_logger::init_once(
android_logger::Config::default()
.with_tag("keymint-hal-trusty")
- .with_min_level(log::Level::Info)
- .with_log_id(android_logger::LogId::System),
+ .with_max_level(log::LevelFilter::Info)
+ .with_log_buffer(android_logger::LogId::System),
);
// Redirect panic messages to logcat.
panic::set_hook(Box::new(|panic_info| {
diff --git a/trusty/metrics/include/trusty/metrics/tipc.h b/trusty/metrics/include/trusty/metrics/tipc.h
index 66d0876..b29628e 100644
--- a/trusty/metrics/include/trusty/metrics/tipc.h
+++ b/trusty/metrics/include/trusty/metrics/tipc.h
@@ -91,9 +91,12 @@
* struct metrics_report_crash_req - arguments of %METRICS_CMD_REPORT_CRASH
* requests
* @app_id_len: length of app ID that follows this structure
+ * @crash_reason: architecture-specific code representing the reason for the
+ * crash
*/
struct metrics_report_crash_req {
uint32_t app_id_len;
+ uint32_t crash_reason;
} __attribute__((__packed__));
#define METRICS_MAX_APP_ID_LEN 256
diff --git a/trusty/secretkeeper/src/hal_main.rs b/trusty/secretkeeper/src/hal_main.rs
index 1dc697d..df30493 100644
--- a/trusty/secretkeeper/src/hal_main.rs
+++ b/trusty/secretkeeper/src/hal_main.rs
@@ -101,8 +101,8 @@
android_logger::init_once(
android_logger::Config::default()
.with_tag("secretkeeper-hal-trusty")
- .with_min_level(log::Level::Info)
- .with_log_id(android_logger::LogId::System),
+ .with_max_level(log::LevelFilter::Info)
+ .with_log_buffer(android_logger::LogId::System),
);
// Redirect panic messages to logcat.
panic::set_hook(Box::new(|panic_info| {
diff --git a/trusty/trusty-base.mk b/trusty/trusty-base.mk
index 5aa4392..b21eca6 100644
--- a/trusty/trusty-base.mk
+++ b/trusty/trusty-base.mk
@@ -35,7 +35,6 @@
LOCAL_KEYMINT_PRODUCT_PACKAGE := android.hardware.security.keymint-service.trusty
endif
-# TODO(b/306364873): move this to be flag-controlled?
ifeq ($(SECRETKEEPER_ENABLED),true)
LOCAL_SECRETKEEPER_PRODUCT_PACKAGE := android.hardware.security.secretkeeper.trusty
else