From 52eee940ed5cf15eb9f6839697367d5cad847118 Mon Sep 17 00:00:00 2001 From: Steven Moreland Date: Thu, 3 Jun 2021 00:59:28 +0000 Subject: libbinder: RPC prevent oneway cmd buildup Previously, binderRpcTest OnewayStressTest was having a flaky hang @ 100 calls because the sockets were filling up with decref counts (these are currently a bit long because they contain RpcAddress - these may need to be smaller - see b/182939933). I temporarily reduced this to 50 calls to avoid the flake, but this is the real fix (which also increases the number of calls to 500). Now, when we send a oneway call, we also drain any refcounts ("CONTROL") commands. We disallow processing of any transactions in this case (this would indicate a misbehaving client). Fixes: 189272263 Test: binderRpcTest Change-Id: If92af65f5e5a920e39046af6b10f675b5c26cd6e --- libs/binder/RpcState.cpp | 27 ++++++++++++++++++++++----- 1 file changed, 22 insertions(+), 5 deletions(-) (limited to 'libs/binder/RpcState.cpp') diff --git a/libs/binder/RpcState.cpp b/libs/binder/RpcState.cpp index e18179e4a8..f1cbe0d4a4 100644 --- a/libs/binder/RpcState.cpp +++ b/libs/binder/RpcState.cpp @@ -386,7 +386,11 @@ status_t RpcState::transactAddress(const base::unique_fd& fd, const RpcAddress& if (flags & IBinder::FLAG_ONEWAY) { LOG_RPC_DETAIL("Oneway command, so no longer waiting on %d", fd.get()); - return OK; // do not wait for result + + // Do not wait on result. + // However, too many oneway calls may cause refcounts to build up and fill up the socket, + // so process those. + return drainCommands(fd, session, CommandType::CONTROL_ONLY); } LOG_ALWAYS_FATAL_IF(reply == nullptr, "Reply parcel must be used for synchronous transaction."); @@ -413,7 +417,8 @@ status_t RpcState::waitForReply(const base::unique_fd& fd, const sp& if (command.command == RPC_COMMAND_REPLY) break; - if (status_t status = processServerCommand(fd, session, command); status != OK) + if (status_t status = processServerCommand(fd, session, command, CommandType::ANY); + status != OK) return status; } @@ -471,7 +476,8 @@ status_t RpcState::sendDecStrong(const base::unique_fd& fd, const RpcAddress& ad return OK; } -status_t RpcState::getAndExecuteCommand(const base::unique_fd& fd, const sp& session) { +status_t RpcState::getAndExecuteCommand(const base::unique_fd& fd, const sp& session, + CommandType type) { LOG_RPC_DETAIL("getAndExecuteCommand on fd %d", fd.get()); RpcWireHeader command; @@ -479,11 +485,21 @@ status_t RpcState::getAndExecuteCommand(const base::unique_fd& fd, const sp& session, + CommandType type) { + uint8_t buf; + while (0 < TEMP_FAILURE_RETRY(recv(fd.get(), &buf, sizeof(buf), MSG_PEEK | MSG_DONTWAIT))) { + status_t status = getAndExecuteCommand(fd, session, type); + if (status != OK) return status; + } + return OK; } status_t RpcState::processServerCommand(const base::unique_fd& fd, const sp& session, - const RpcWireHeader& command) { + const RpcWireHeader& command, CommandType type) { IPCThreadState* kernelBinderState = IPCThreadState::selfOrNull(); IPCThreadState::SpGuard spGuard{ .address = __builtin_frame_address(0), @@ -501,6 +517,7 @@ status_t RpcState::processServerCommand(const base::unique_fd& fd, const sp