| /* | 
 |  * Copyright (C) 2020 The Android Open Source Project | 
 |  * | 
 |  * Licensed under the Apache License, Version 2.0 (the "License"); | 
 |  * you may not use this file except in compliance with the License. | 
 |  * You may obtain a copy of the License at | 
 |  * | 
 |  *      http://www.apache.org/licenses/LICENSE-2.0 | 
 |  * | 
 |  * Unless required by applicable law or agreed to in writing, software | 
 |  * distributed under the License is distributed on an "AS IS" BASIS, | 
 |  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
 |  * See the License for the specific language governing permissions and | 
 |  * limitations under the License. | 
 |  */ | 
 |  | 
 | #include <android-base/stringprintf.h> | 
 | #include <gtest/gtest.h> | 
 |  | 
 | #include <chrono> | 
 | #include <cstdlib> | 
 | #include <iostream> | 
 | #include <thread> | 
 | #include <type_traits> | 
 |  | 
 | #include <dlfcn.h> | 
 | #include <poll.h> | 
 | #include <sys/prctl.h> | 
 | #include <sys/socket.h> | 
 |  | 
 | #include "binderRpcTestCommon.h" | 
 |  | 
 | using namespace std::chrono_literals; | 
 | using namespace std::placeholders; | 
 | using testing::AssertionFailure; | 
 | using testing::AssertionResult; | 
 | using testing::AssertionSuccess; | 
 |  | 
 | namespace android { | 
 |  | 
 | #ifdef BINDER_TEST_NO_SHARED_LIBS | 
 | constexpr bool kEnableSharedLibs = false; | 
 | #else | 
 | constexpr bool kEnableSharedLibs = true; | 
 | #endif | 
 |  | 
 | static_assert(RPC_WIRE_PROTOCOL_VERSION + 1 == RPC_WIRE_PROTOCOL_VERSION_NEXT || | 
 |               RPC_WIRE_PROTOCOL_VERSION == RPC_WIRE_PROTOCOL_VERSION_EXPERIMENTAL); | 
 |  | 
 | TEST(BinderRpcParcel, EntireParcelFormatted) { | 
 |     Parcel p; | 
 |     p.writeInt32(3); | 
 |  | 
 |     EXPECT_DEATH(p.markForBinder(sp<BBinder>::make()), "format must be set before data is written"); | 
 | } | 
 |  | 
 | TEST(BinderRpc, CannotUseNextWireVersion) { | 
 |     auto session = RpcSession::make(); | 
 |     EXPECT_FALSE(session->setProtocolVersion(RPC_WIRE_PROTOCOL_VERSION_NEXT)); | 
 |     EXPECT_FALSE(session->setProtocolVersion(RPC_WIRE_PROTOCOL_VERSION_NEXT + 1)); | 
 |     EXPECT_FALSE(session->setProtocolVersion(RPC_WIRE_PROTOCOL_VERSION_NEXT + 2)); | 
 |     EXPECT_FALSE(session->setProtocolVersion(RPC_WIRE_PROTOCOL_VERSION_NEXT + 15)); | 
 | } | 
 |  | 
 | TEST(BinderRpc, CanUseExperimentalWireVersion) { | 
 |     auto session = RpcSession::make(); | 
 |     EXPECT_TRUE(session->setProtocolVersion(RPC_WIRE_PROTOCOL_VERSION_EXPERIMENTAL)); | 
 | } | 
 |  | 
 | using android::binder::Status; | 
 |  | 
 | #define EXPECT_OK(status)                 \ | 
 |     do {                                  \ | 
 |         Status stat = (status);           \ | 
 |         EXPECT_TRUE(stat.isOk()) << stat; \ | 
 |     } while (false) | 
 |  | 
 | static std::string WaitStatusToString(int wstatus) { | 
 |     if (WIFEXITED(wstatus)) { | 
 |         return base::StringPrintf("exit status %d", WEXITSTATUS(wstatus)); | 
 |     } | 
 |     if (WIFSIGNALED(wstatus)) { | 
 |         return base::StringPrintf("term signal %d", WTERMSIG(wstatus)); | 
 |     } | 
 |     return base::StringPrintf("unexpected state %d", wstatus); | 
 | } | 
 |  | 
 | static void debugBacktrace(pid_t pid) { | 
 |     std::cerr << "TAKING BACKTRACE FOR PID " << pid << std::endl; | 
 |     system((std::string("debuggerd -b ") + std::to_string(pid)).c_str()); | 
 | } | 
 |  | 
 | class Process { | 
 | public: | 
 |     Process(Process&&) = default; | 
 |     Process(const std::function<void(android::base::borrowed_fd /* writeEnd */, | 
 |                                      android::base::borrowed_fd /* readEnd */)>& f) { | 
 |         android::base::unique_fd childWriteEnd; | 
 |         android::base::unique_fd childReadEnd; | 
 |         CHECK(android::base::Pipe(&mReadEnd, &childWriteEnd, 0)) << strerror(errno); | 
 |         CHECK(android::base::Pipe(&childReadEnd, &mWriteEnd, 0)) << strerror(errno); | 
 |         if (0 == (mPid = fork())) { | 
 |             // racey: assume parent doesn't crash before this is set | 
 |             prctl(PR_SET_PDEATHSIG, SIGHUP); | 
 |  | 
 |             f(childWriteEnd, childReadEnd); | 
 |  | 
 |             exit(0); | 
 |         } | 
 |     } | 
 |     ~Process() { | 
 |         if (mPid != 0) { | 
 |             int wstatus; | 
 |             waitpid(mPid, &wstatus, 0); | 
 |             if (mCustomExitStatusCheck) { | 
 |                 mCustomExitStatusCheck(wstatus); | 
 |             } else { | 
 |                 EXPECT_TRUE(WIFEXITED(wstatus) && WEXITSTATUS(wstatus) == 0) | 
 |                         << "server process failed: " << WaitStatusToString(wstatus); | 
 |             } | 
 |         } | 
 |     } | 
 |     android::base::borrowed_fd readEnd() { return mReadEnd; } | 
 |     android::base::borrowed_fd writeEnd() { return mWriteEnd; } | 
 |  | 
 |     void setCustomExitStatusCheck(std::function<void(int wstatus)> f) { | 
 |         mCustomExitStatusCheck = std::move(f); | 
 |     } | 
 |  | 
 |     // Kill the process. Avoid if possible. Shutdown gracefully via an RPC instead. | 
 |     void terminate() { kill(mPid, SIGTERM); } | 
 |  | 
 |     pid_t getPid() { return mPid; } | 
 |  | 
 | private: | 
 |     std::function<void(int wstatus)> mCustomExitStatusCheck; | 
 |     pid_t mPid = 0; | 
 |     android::base::unique_fd mReadEnd; | 
 |     android::base::unique_fd mWriteEnd; | 
 | }; | 
 |  | 
 | static std::string allocateSocketAddress() { | 
 |     static size_t id = 0; | 
 |     std::string temp = getenv("TMPDIR") ?: "/tmp"; | 
 |     auto ret = temp + "/binderRpcTest_" + std::to_string(id++); | 
 |     unlink(ret.c_str()); | 
 |     return ret; | 
 | }; | 
 |  | 
 | static unsigned int allocateVsockPort() { | 
 |     static unsigned int vsockPort = 34567; | 
 |     return vsockPort++; | 
 | } | 
 |  | 
 | struct ProcessSession { | 
 |     // reference to process hosting a socket server | 
 |     Process host; | 
 |  | 
 |     struct SessionInfo { | 
 |         sp<RpcSession> session; | 
 |         sp<IBinder> root; | 
 |     }; | 
 |  | 
 |     // client session objects associated with other process | 
 |     // each one represents a separate session | 
 |     std::vector<SessionInfo> sessions; | 
 |  | 
 |     ProcessSession(ProcessSession&&) = default; | 
 |     ~ProcessSession() { | 
 |         for (auto& session : sessions) { | 
 |             session.root = nullptr; | 
 |         } | 
 |  | 
 |         for (auto& info : sessions) { | 
 |             sp<RpcSession>& session = info.session; | 
 |  | 
 |             EXPECT_NE(nullptr, session); | 
 |             EXPECT_NE(nullptr, session->state()); | 
 |             EXPECT_EQ(0, session->state()->countBinders()) << (session->state()->dump(), "dump:"); | 
 |  | 
 |             wp<RpcSession> weakSession = session; | 
 |             session = nullptr; | 
 |  | 
 |             // b/244325464 - 'getStrongCount' is printing '1' on failure here, which indicates the | 
 |             // the object should not actually be promotable. By looping, we distinguish a race here | 
 |             // from a bug causing the object to not be promotable. | 
 |             for (size_t i = 0; i < 3; i++) { | 
 |                 sp<RpcSession> strongSession = weakSession.promote(); | 
 |                 EXPECT_EQ(nullptr, strongSession) | 
 |                         << (debugBacktrace(host.getPid()), debugBacktrace(getpid()), | 
 |                             "Leaked sess: ") | 
 |                         << strongSession->getStrongCount() << " checked time " << i; | 
 |  | 
 |                 if (strongSession != nullptr) { | 
 |                     sleep(1); | 
 |                 } | 
 |             } | 
 |         } | 
 |     } | 
 | }; | 
 |  | 
 | // Process session where the process hosts IBinderRpcTest, the server used | 
 | // for most testing here | 
 | struct BinderRpcTestProcessSession { | 
 |     ProcessSession proc; | 
 |  | 
 |     // pre-fetched root object (for first session) | 
 |     sp<IBinder> rootBinder; | 
 |  | 
 |     // pre-casted root object (for first session) | 
 |     sp<IBinderRpcTest> rootIface; | 
 |  | 
 |     // whether session should be invalidated by end of run | 
 |     bool expectAlreadyShutdown = false; | 
 |  | 
 |     BinderRpcTestProcessSession(BinderRpcTestProcessSession&&) = default; | 
 |     ~BinderRpcTestProcessSession() { | 
 |         if (!expectAlreadyShutdown) { | 
 |             EXPECT_NE(nullptr, rootIface); | 
 |             if (rootIface == nullptr) return; | 
 |  | 
 |             std::vector<int32_t> remoteCounts; | 
 |             // calling over any sessions counts across all sessions | 
 |             EXPECT_OK(rootIface->countBinders(&remoteCounts)); | 
 |             EXPECT_EQ(remoteCounts.size(), proc.sessions.size()); | 
 |             for (auto remoteCount : remoteCounts) { | 
 |                 EXPECT_EQ(remoteCount, 1); | 
 |             } | 
 |  | 
 |             // even though it is on another thread, shutdown races with | 
 |             // the transaction reply being written | 
 |             if (auto status = rootIface->scheduleShutdown(); !status.isOk()) { | 
 |                 EXPECT_EQ(DEAD_OBJECT, status.transactionError()) << status; | 
 |             } | 
 |         } | 
 |  | 
 |         rootIface = nullptr; | 
 |         rootBinder = nullptr; | 
 |     } | 
 | }; | 
 |  | 
 | static base::unique_fd connectTo(const RpcSocketAddress& addr) { | 
 |     base::unique_fd serverFd( | 
 |             TEMP_FAILURE_RETRY(socket(addr.addr()->sa_family, SOCK_STREAM | SOCK_CLOEXEC, 0))); | 
 |     int savedErrno = errno; | 
 |     CHECK(serverFd.ok()) << "Could not create socket " << addr.toString() << ": " | 
 |                          << strerror(savedErrno); | 
 |  | 
 |     if (0 != TEMP_FAILURE_RETRY(connect(serverFd.get(), addr.addr(), addr.addrSize()))) { | 
 |         int savedErrno = errno; | 
 |         LOG(FATAL) << "Could not connect to socket " << addr.toString() << ": " | 
 |                    << strerror(savedErrno); | 
 |     } | 
 |     return serverFd; | 
 | } | 
 |  | 
 | static base::unique_fd connectToUnixBootstrap(const RpcTransportFd& transportFd) { | 
 |     base::unique_fd sockClient, sockServer; | 
 |     if (!base::Socketpair(SOCK_STREAM, &sockClient, &sockServer)) { | 
 |         int savedErrno = errno; | 
 |         LOG(FATAL) << "Failed socketpair(): " << strerror(savedErrno); | 
 |     } | 
 |  | 
 |     int zero = 0; | 
 |     iovec iov{&zero, sizeof(zero)}; | 
 |     std::vector<std::variant<base::unique_fd, base::borrowed_fd>> fds; | 
 |     fds.emplace_back(std::move(sockServer)); | 
 |  | 
 |     if (sendMessageOnSocket(transportFd, &iov, 1, &fds) < 0) { | 
 |         int savedErrno = errno; | 
 |         LOG(FATAL) << "Failed sendMessageOnSocket: " << strerror(savedErrno); | 
 |     } | 
 |     return std::move(sockClient); | 
 | } | 
 |  | 
 | using RunServiceFn = void (*)(android::base::borrowed_fd writeEnd, | 
 |                               android::base::borrowed_fd readEnd); | 
 |  | 
 | class BinderRpc : public ::testing::TestWithParam< | 
 |                           std::tuple<SocketType, RpcSecurity, uint32_t, uint32_t, bool, bool>> { | 
 | public: | 
 |     SocketType socketType() const { return std::get<0>(GetParam()); } | 
 |     RpcSecurity rpcSecurity() const { return std::get<1>(GetParam()); } | 
 |     uint32_t clientVersion() const { return std::get<2>(GetParam()); } | 
 |     uint32_t serverVersion() const { return std::get<3>(GetParam()); } | 
 |     bool serverSingleThreaded() const { return std::get<4>(GetParam()); } | 
 |     bool noKernel() const { return std::get<5>(GetParam()); } | 
 |  | 
 |     bool clientOrServerSingleThreaded() const { | 
 |         return !kEnableRpcThreads || serverSingleThreaded(); | 
 |     } | 
 |  | 
 |     // Whether the test params support sending FDs in parcels. | 
 |     bool supportsFdTransport() const { | 
 |         return clientVersion() >= 1 && serverVersion() >= 1 && rpcSecurity() != RpcSecurity::TLS && | 
 |                 (socketType() == SocketType::PRECONNECTED || socketType() == SocketType::UNIX || | 
 |                  socketType() == SocketType::UNIX_BOOTSTRAP); | 
 |     } | 
 |  | 
 |     void SetUp() override { | 
 |         if (socketType() == SocketType::UNIX_BOOTSTRAP && rpcSecurity() == RpcSecurity::TLS) { | 
 |             GTEST_SKIP() << "Unix bootstrap not supported over a TLS transport"; | 
 |         } | 
 |     } | 
 |  | 
 |     static inline std::string PrintParamInfo(const testing::TestParamInfo<ParamType>& info) { | 
 |         auto [type, security, clientVersion, serverVersion, singleThreaded, noKernel] = info.param; | 
 |         auto ret = PrintToString(type) + "_" + newFactory(security)->toCString() + "_clientV" + | 
 |                 std::to_string(clientVersion) + "_serverV" + std::to_string(serverVersion); | 
 |         if (singleThreaded) { | 
 |             ret += "_single_threaded"; | 
 |         } | 
 |         if (noKernel) { | 
 |             ret += "_no_kernel"; | 
 |         } | 
 |         return ret; | 
 |     } | 
 |  | 
 |     // This creates a new process serving an interface on a certain number of | 
 |     // threads. | 
 |     ProcessSession createRpcTestSocketServerProcessEtc(const BinderRpcOptions& options) { | 
 |         CHECK_GE(options.numSessions, 1) << "Must have at least one session to a server"; | 
 |  | 
 |         SocketType socketType = std::get<0>(GetParam()); | 
 |         RpcSecurity rpcSecurity = std::get<1>(GetParam()); | 
 |         uint32_t clientVersion = std::get<2>(GetParam()); | 
 |         uint32_t serverVersion = std::get<3>(GetParam()); | 
 |         bool singleThreaded = std::get<4>(GetParam()); | 
 |         bool noKernel = std::get<5>(GetParam()); | 
 |  | 
 |         std::string path = android::base::GetExecutableDirectory(); | 
 |         auto servicePath = | 
 |                 android::base::StringPrintf("%s/binder_rpc_test_service%s%s", path.c_str(), | 
 |                                             singleThreaded ? "_single_threaded" : "", | 
 |                                             noKernel ? "_no_kernel" : ""); | 
 |  | 
 |         base::unique_fd bootstrapClientFd, bootstrapServerFd; | 
 |         // Do not set O_CLOEXEC, bootstrapServerFd needs to survive fork/exec. | 
 |         // This is because we cannot pass ParcelFileDescriptor over a pipe. | 
 |         if (!base::Socketpair(SOCK_STREAM, &bootstrapClientFd, &bootstrapServerFd)) { | 
 |             int savedErrno = errno; | 
 |             LOG(FATAL) << "Failed socketpair(): " << strerror(savedErrno); | 
 |         } | 
 |  | 
 |         auto ret = ProcessSession{ | 
 |                 .host = Process([=](android::base::borrowed_fd writeEnd, | 
 |                                     android::base::borrowed_fd readEnd) { | 
 |                     auto writeFd = std::to_string(writeEnd.get()); | 
 |                     auto readFd = std::to_string(readEnd.get()); | 
 |                     execl(servicePath.c_str(), servicePath.c_str(), writeFd.c_str(), readFd.c_str(), | 
 |                           NULL); | 
 |                 }), | 
 |         }; | 
 |  | 
 |         BinderRpcTestServerConfig serverConfig; | 
 |         serverConfig.numThreads = options.numThreads; | 
 |         serverConfig.socketType = static_cast<int32_t>(socketType); | 
 |         serverConfig.rpcSecurity = static_cast<int32_t>(rpcSecurity); | 
 |         serverConfig.serverVersion = serverVersion; | 
 |         serverConfig.vsockPort = allocateVsockPort(); | 
 |         serverConfig.addr = allocateSocketAddress(); | 
 |         serverConfig.unixBootstrapFd = bootstrapServerFd.get(); | 
 |         for (auto mode : options.serverSupportedFileDescriptorTransportModes) { | 
 |             serverConfig.serverSupportedFileDescriptorTransportModes.push_back( | 
 |                     static_cast<int32_t>(mode)); | 
 |         } | 
 |         writeToFd(ret.host.writeEnd(), serverConfig); | 
 |  | 
 |         std::vector<sp<RpcSession>> sessions; | 
 |         auto certVerifier = std::make_shared<RpcCertificateVerifierSimple>(); | 
 |         for (size_t i = 0; i < options.numSessions; i++) { | 
 |             sessions.emplace_back(RpcSession::make(newFactory(rpcSecurity, certVerifier))); | 
 |         } | 
 |  | 
 |         auto serverInfo = readFromFd<BinderRpcTestServerInfo>(ret.host.readEnd()); | 
 |         BinderRpcTestClientInfo clientInfo; | 
 |         for (const auto& session : sessions) { | 
 |             auto& parcelableCert = clientInfo.certs.emplace_back(); | 
 |             parcelableCert.data = session->getCertificate(RpcCertificateFormat::PEM); | 
 |         } | 
 |         writeToFd(ret.host.writeEnd(), clientInfo); | 
 |  | 
 |         CHECK_LE(serverInfo.port, std::numeric_limits<unsigned int>::max()); | 
 |         if (socketType == SocketType::INET) { | 
 |             CHECK_NE(0, serverInfo.port); | 
 |         } | 
 |  | 
 |         if (rpcSecurity == RpcSecurity::TLS) { | 
 |             const auto& serverCert = serverInfo.cert.data; | 
 |             CHECK_EQ(OK, | 
 |                      certVerifier->addTrustedPeerCertificate(RpcCertificateFormat::PEM, | 
 |                                                              serverCert)); | 
 |         } | 
 |  | 
 |         status_t status; | 
 |  | 
 |         for (const auto& session : sessions) { | 
 |             CHECK(session->setProtocolVersion(clientVersion)); | 
 |             session->setMaxIncomingThreads(options.numIncomingConnections); | 
 |             session->setMaxOutgoingThreads(options.numOutgoingConnections); | 
 |             session->setFileDescriptorTransportMode(options.clientFileDescriptorTransportMode); | 
 |  | 
 |             switch (socketType) { | 
 |                 case SocketType::PRECONNECTED: | 
 |                     status = session->setupPreconnectedClient({}, [=]() { | 
 |                         return connectTo(UnixSocketAddress(serverConfig.addr.c_str())); | 
 |                     }); | 
 |                     break; | 
 |                 case SocketType::UNIX: | 
 |                     status = session->setupUnixDomainClient(serverConfig.addr.c_str()); | 
 |                     break; | 
 |                 case SocketType::UNIX_BOOTSTRAP: | 
 |                     status = session->setupUnixDomainSocketBootstrapClient( | 
 |                             base::unique_fd(dup(bootstrapClientFd.get()))); | 
 |                     break; | 
 |                 case SocketType::VSOCK: | 
 |                     status = session->setupVsockClient(VMADDR_CID_LOCAL, serverConfig.vsockPort); | 
 |                     break; | 
 |                 case SocketType::INET: | 
 |                     status = session->setupInetClient("127.0.0.1", serverInfo.port); | 
 |                     break; | 
 |                 default: | 
 |                     LOG_ALWAYS_FATAL("Unknown socket type"); | 
 |             } | 
 |             if (options.allowConnectFailure && status != OK) { | 
 |                 ret.sessions.clear(); | 
 |                 break; | 
 |             } | 
 |             CHECK_EQ(status, OK) << "Could not connect: " << statusToString(status); | 
 |             ret.sessions.push_back({session, session->getRootObject()}); | 
 |         } | 
 |         return ret; | 
 |     } | 
 |  | 
 |     BinderRpcTestProcessSession createRpcTestSocketServerProcess(const BinderRpcOptions& options) { | 
 |         BinderRpcTestProcessSession ret{ | 
 |                 .proc = createRpcTestSocketServerProcessEtc(options), | 
 |         }; | 
 |  | 
 |         ret.rootBinder = ret.proc.sessions.empty() ? nullptr : ret.proc.sessions.at(0).root; | 
 |         ret.rootIface = interface_cast<IBinderRpcTest>(ret.rootBinder); | 
 |  | 
 |         return ret; | 
 |     } | 
 |  | 
 |     void testThreadPoolOverSaturated(sp<IBinderRpcTest> iface, size_t numCalls, | 
 |                                      size_t sleepMs = 500); | 
 | }; | 
 |  | 
 | TEST_P(BinderRpc, Ping) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     ASSERT_NE(proc.rootBinder, nullptr); | 
 |     EXPECT_EQ(OK, proc.rootBinder->pingBinder()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, GetInterfaceDescriptor) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     ASSERT_NE(proc.rootBinder, nullptr); | 
 |     EXPECT_EQ(IBinderRpcTest::descriptor, proc.rootBinder->getInterfaceDescriptor()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, MultipleSessions) { | 
 |     if (serverSingleThreaded()) { | 
 |         // Tests with multiple sessions require a multi-threaded service, | 
 |         // but work fine on a single-threaded client | 
 |         GTEST_SKIP() << "This test requires a multi-threaded service"; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({.numThreads = 1, .numSessions = 5}); | 
 |     for (auto session : proc.proc.sessions) { | 
 |         ASSERT_NE(nullptr, session.root); | 
 |         EXPECT_EQ(OK, session.root->pingBinder()); | 
 |     } | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SeparateRootObject) { | 
 |     if (serverSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires a multi-threaded service"; | 
 |     } | 
 |  | 
 |     SocketType type = std::get<0>(GetParam()); | 
 |     if (type == SocketType::PRECONNECTED || type == SocketType::UNIX || | 
 |         type == SocketType::UNIX_BOOTSTRAP) { | 
 |         // we can't get port numbers for unix sockets | 
 |         return; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({.numSessions = 2}); | 
 |  | 
 |     int port1 = 0; | 
 |     EXPECT_OK(proc.rootIface->getClientPort(&port1)); | 
 |  | 
 |     sp<IBinderRpcTest> rootIface2 = interface_cast<IBinderRpcTest>(proc.proc.sessions.at(1).root); | 
 |     int port2; | 
 |     EXPECT_OK(rootIface2->getClientPort(&port2)); | 
 |  | 
 |     // we should have a different IBinderRpcTest object created for each | 
 |     // session, because we use setPerSessionRootObject | 
 |     EXPECT_NE(port1, port2); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, TransactionsMustBeMarkedRpc) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     Parcel data; | 
 |     Parcel reply; | 
 |     EXPECT_EQ(BAD_TYPE, proc.rootBinder->transact(IBinder::PING_TRANSACTION, data, &reply, 0)); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, AppendSeparateFormats) { | 
 |     auto proc1 = createRpcTestSocketServerProcess({}); | 
 |     auto proc2 = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     Parcel pRaw; | 
 |  | 
 |     Parcel p1; | 
 |     p1.markForBinder(proc1.rootBinder); | 
 |     p1.writeInt32(3); | 
 |  | 
 |     EXPECT_EQ(BAD_TYPE, p1.appendFrom(&pRaw, 0, pRaw.dataSize())); | 
 |     EXPECT_EQ(BAD_TYPE, pRaw.appendFrom(&p1, 0, p1.dataSize())); | 
 |  | 
 |     Parcel p2; | 
 |     p2.markForBinder(proc2.rootBinder); | 
 |     p2.writeInt32(7); | 
 |  | 
 |     EXPECT_EQ(BAD_TYPE, p1.appendFrom(&p2, 0, p2.dataSize())); | 
 |     EXPECT_EQ(BAD_TYPE, p2.appendFrom(&p1, 0, p1.dataSize())); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, UnknownTransaction) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     Parcel data; | 
 |     data.markForBinder(proc.rootBinder); | 
 |     Parcel reply; | 
 |     EXPECT_EQ(UNKNOWN_TRANSACTION, proc.rootBinder->transact(1337, data, &reply, 0)); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SendSomethingOneway) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     EXPECT_OK(proc.rootIface->sendString("asdf")); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SendAndGetResultBack) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     std::string doubled; | 
 |     EXPECT_OK(proc.rootIface->doubleString("cool ", &doubled)); | 
 |     EXPECT_EQ("cool cool ", doubled); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SendAndGetResultBackBig) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     std::string single = std::string(1024, 'a'); | 
 |     std::string doubled; | 
 |     EXPECT_OK(proc.rootIface->doubleString(single, &doubled)); | 
 |     EXPECT_EQ(single + single, doubled); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, InvalidNullBinderReturn) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinder> outBinder; | 
 |     EXPECT_EQ(proc.rootIface->getNullBinder(&outBinder).transactionError(), UNEXPECTED_NULL); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, CallMeBack) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     int32_t pingResult; | 
 |     EXPECT_OK(proc.rootIface->pingMe(new MyBinderRpcSession("foo"), &pingResult)); | 
 |     EXPECT_EQ(OK, pingResult); | 
 |  | 
 |     EXPECT_EQ(0, MyBinderRpcSession::gNum); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, RepeatBinder) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinder> inBinder = new MyBinderRpcSession("foo"); | 
 |     sp<IBinder> outBinder; | 
 |     EXPECT_OK(proc.rootIface->repeatBinder(inBinder, &outBinder)); | 
 |     EXPECT_EQ(inBinder, outBinder); | 
 |  | 
 |     wp<IBinder> weak = inBinder; | 
 |     inBinder = nullptr; | 
 |     outBinder = nullptr; | 
 |  | 
 |     // Force reading a reply, to process any pending dec refs from the other | 
 |     // process (the other process will process dec refs there before processing | 
 |     // the ping here). | 
 |     EXPECT_EQ(OK, proc.rootBinder->pingBinder()); | 
 |  | 
 |     EXPECT_EQ(nullptr, weak.promote()); | 
 |  | 
 |     EXPECT_EQ(0, MyBinderRpcSession::gNum); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, RepeatTheirBinder) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinderRpcSession> session; | 
 |     EXPECT_OK(proc.rootIface->openSession("aoeu", &session)); | 
 |  | 
 |     sp<IBinder> inBinder = IInterface::asBinder(session); | 
 |     sp<IBinder> outBinder; | 
 |     EXPECT_OK(proc.rootIface->repeatBinder(inBinder, &outBinder)); | 
 |     EXPECT_EQ(inBinder, outBinder); | 
 |  | 
 |     wp<IBinder> weak = inBinder; | 
 |     session = nullptr; | 
 |     inBinder = nullptr; | 
 |     outBinder = nullptr; | 
 |  | 
 |     // Force reading a reply, to process any pending dec refs from the other | 
 |     // process (the other process will process dec refs there before processing | 
 |     // the ping here). | 
 |     EXPECT_EQ(OK, proc.rootBinder->pingBinder()); | 
 |  | 
 |     EXPECT_EQ(nullptr, weak.promote()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, RepeatBinderNull) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinder> outBinder; | 
 |     EXPECT_OK(proc.rootIface->repeatBinder(nullptr, &outBinder)); | 
 |     EXPECT_EQ(nullptr, outBinder); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, HoldBinder) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     IBinder* ptr = nullptr; | 
 |     { | 
 |         sp<IBinder> binder = new BBinder(); | 
 |         ptr = binder.get(); | 
 |         EXPECT_OK(proc.rootIface->holdBinder(binder)); | 
 |     } | 
 |  | 
 |     sp<IBinder> held; | 
 |     EXPECT_OK(proc.rootIface->getHeldBinder(&held)); | 
 |  | 
 |     EXPECT_EQ(held.get(), ptr); | 
 |  | 
 |     // stop holding binder, because we test to make sure references are cleaned | 
 |     // up | 
 |     EXPECT_OK(proc.rootIface->holdBinder(nullptr)); | 
 |     // and flush ref counts | 
 |     EXPECT_EQ(OK, proc.rootBinder->pingBinder()); | 
 | } | 
 |  | 
 | // START TESTS FOR LIMITATIONS OF SOCKET BINDER | 
 | // These are behavioral differences form regular binder, where certain usecases | 
 | // aren't supported. | 
 |  | 
 | TEST_P(BinderRpc, CannotMixBindersBetweenUnrelatedSocketSessions) { | 
 |     auto proc1 = createRpcTestSocketServerProcess({}); | 
 |     auto proc2 = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinder> outBinder; | 
 |     EXPECT_EQ(INVALID_OPERATION, | 
 |               proc1.rootIface->repeatBinder(proc2.rootBinder, &outBinder).transactionError()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, CannotMixBindersBetweenTwoSessionsToTheSameServer) { | 
 |     if (serverSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires a multi-threaded service"; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({.numThreads = 1, .numSessions = 2}); | 
 |  | 
 |     sp<IBinder> outBinder; | 
 |     EXPECT_EQ(INVALID_OPERATION, | 
 |               proc.rootIface->repeatBinder(proc.proc.sessions.at(1).root, &outBinder) | 
 |                       .transactionError()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, CannotSendRegularBinderOverSocketBinder) { | 
 |     if (!kEnableKernelIpc || noKernel()) { | 
 |         GTEST_SKIP() << "Test disabled because Binder kernel driver was disabled " | 
 |                         "at build time."; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinder> someRealBinder = IInterface::asBinder(defaultServiceManager()); | 
 |     sp<IBinder> outBinder; | 
 |     EXPECT_EQ(INVALID_OPERATION, | 
 |               proc.rootIface->repeatBinder(someRealBinder, &outBinder).transactionError()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, CannotSendSocketBinderOverRegularBinder) { | 
 |     if (!kEnableKernelIpc || noKernel()) { | 
 |         GTEST_SKIP() << "Test disabled because Binder kernel driver was disabled " | 
 |                         "at build time."; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     // for historical reasons, IServiceManager interface only returns the | 
 |     // exception code | 
 |     EXPECT_EQ(binder::Status::EX_TRANSACTION_FAILED, | 
 |               defaultServiceManager()->addService(String16("not_suspicious"), proc.rootBinder)); | 
 | } | 
 |  | 
 | // END TESTS FOR LIMITATIONS OF SOCKET BINDER | 
 |  | 
 | TEST_P(BinderRpc, RepeatRootObject) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinder> outBinder; | 
 |     EXPECT_OK(proc.rootIface->repeatBinder(proc.rootBinder, &outBinder)); | 
 |     EXPECT_EQ(proc.rootBinder, outBinder); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, NestedTransactions) { | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             // Enable FD support because it uses more stack space and so represents | 
 |             // something closer to a worst case scenario. | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::UNIX, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::UNIX}, | 
 |     }); | 
 |  | 
 |     auto nastyNester = sp<MyBinderRpcTest>::make(); | 
 |     EXPECT_OK(proc.rootIface->nestMe(nastyNester, 10)); | 
 |  | 
 |     wp<IBinder> weak = nastyNester; | 
 |     nastyNester = nullptr; | 
 |     EXPECT_EQ(nullptr, weak.promote()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SameBinderEquality) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinder> a; | 
 |     EXPECT_OK(proc.rootIface->alwaysGiveMeTheSameBinder(&a)); | 
 |  | 
 |     sp<IBinder> b; | 
 |     EXPECT_OK(proc.rootIface->alwaysGiveMeTheSameBinder(&b)); | 
 |  | 
 |     EXPECT_EQ(a, b); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SameBinderEqualityWeak) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinder> a; | 
 |     EXPECT_OK(proc.rootIface->alwaysGiveMeTheSameBinder(&a)); | 
 |     wp<IBinder> weak = a; | 
 |     a = nullptr; | 
 |  | 
 |     sp<IBinder> b; | 
 |     EXPECT_OK(proc.rootIface->alwaysGiveMeTheSameBinder(&b)); | 
 |  | 
 |     // this is the wrong behavior, since BpBinder | 
 |     // doesn't implement onIncStrongAttempted | 
 |     // but make sure there is no crash | 
 |     EXPECT_EQ(nullptr, weak.promote()); | 
 |  | 
 |     GTEST_SKIP() << "Weak binders aren't currently re-promotable for RPC binder."; | 
 |  | 
 |     // In order to fix this: | 
 |     // - need to have incStrongAttempted reflected across IPC boundary (wait for | 
 |     //   response to promote - round trip...) | 
 |     // - sendOnLastWeakRef, to delete entries out of RpcState table | 
 |     EXPECT_EQ(b, weak.promote()); | 
 | } | 
 |  | 
 | #define expectSessions(expected, iface)                   \ | 
 |     do {                                                  \ | 
 |         int session;                                      \ | 
 |         EXPECT_OK((iface)->getNumOpenSessions(&session)); \ | 
 |         EXPECT_EQ(expected, session);                     \ | 
 |     } while (false) | 
 |  | 
 | TEST_P(BinderRpc, SingleSession) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     sp<IBinderRpcSession> session; | 
 |     EXPECT_OK(proc.rootIface->openSession("aoeu", &session)); | 
 |     std::string out; | 
 |     EXPECT_OK(session->getName(&out)); | 
 |     EXPECT_EQ("aoeu", out); | 
 |  | 
 |     expectSessions(1, proc.rootIface); | 
 |     session = nullptr; | 
 |     expectSessions(0, proc.rootIface); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, ManySessions) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     std::vector<sp<IBinderRpcSession>> sessions; | 
 |  | 
 |     for (size_t i = 0; i < 15; i++) { | 
 |         expectSessions(i, proc.rootIface); | 
 |         sp<IBinderRpcSession> session; | 
 |         EXPECT_OK(proc.rootIface->openSession(std::to_string(i), &session)); | 
 |         sessions.push_back(session); | 
 |     } | 
 |     expectSessions(sessions.size(), proc.rootIface); | 
 |     for (size_t i = 0; i < sessions.size(); i++) { | 
 |         std::string out; | 
 |         EXPECT_OK(sessions.at(i)->getName(&out)); | 
 |         EXPECT_EQ(std::to_string(i), out); | 
 |     } | 
 |     expectSessions(sessions.size(), proc.rootIface); | 
 |  | 
 |     while (!sessions.empty()) { | 
 |         sessions.pop_back(); | 
 |         expectSessions(sessions.size(), proc.rootIface); | 
 |     } | 
 |     expectSessions(0, proc.rootIface); | 
 | } | 
 |  | 
 | size_t epochMillis() { | 
 |     using std::chrono::duration_cast; | 
 |     using std::chrono::milliseconds; | 
 |     using std::chrono::seconds; | 
 |     using std::chrono::system_clock; | 
 |     return duration_cast<milliseconds>(system_clock::now().time_since_epoch()).count(); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, ThreadPoolGreaterThanEqualRequested) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     constexpr size_t kNumThreads = 10; | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({.numThreads = kNumThreads}); | 
 |  | 
 |     EXPECT_OK(proc.rootIface->lock()); | 
 |  | 
 |     // block all but one thread taking locks | 
 |     std::vector<std::thread> ts; | 
 |     for (size_t i = 0; i < kNumThreads - 1; i++) { | 
 |         ts.push_back(std::thread([&] { proc.rootIface->lockUnlock(); })); | 
 |     } | 
 |  | 
 |     usleep(10000); // give chance for calls on other threads | 
 |  | 
 |     // other calls still work | 
 |     EXPECT_EQ(OK, proc.rootBinder->pingBinder()); | 
 |  | 
 |     constexpr size_t blockTimeMs = 50; | 
 |     size_t epochMsBefore = epochMillis(); | 
 |     // after this, we should never see a response within this time | 
 |     EXPECT_OK(proc.rootIface->unlockInMsAsync(blockTimeMs)); | 
 |  | 
 |     // this call should be blocked for blockTimeMs | 
 |     EXPECT_EQ(OK, proc.rootBinder->pingBinder()); | 
 |  | 
 |     size_t epochMsAfter = epochMillis(); | 
 |     EXPECT_GE(epochMsAfter, epochMsBefore + blockTimeMs) << epochMsBefore; | 
 |  | 
 |     for (auto& t : ts) t.join(); | 
 | } | 
 |  | 
 | void BinderRpc::testThreadPoolOverSaturated(sp<IBinderRpcTest> iface, size_t numCalls, | 
 |                                             size_t sleepMs) { | 
 |     size_t epochMsBefore = epochMillis(); | 
 |  | 
 |     std::vector<std::thread> ts; | 
 |     for (size_t i = 0; i < numCalls; i++) { | 
 |         ts.push_back(std::thread([&] { iface->sleepMs(sleepMs); })); | 
 |     } | 
 |  | 
 |     for (auto& t : ts) t.join(); | 
 |  | 
 |     size_t epochMsAfter = epochMillis(); | 
 |  | 
 |     EXPECT_GE(epochMsAfter, epochMsBefore + 2 * sleepMs); | 
 |  | 
 |     // Potential flake, but make sure calls are handled in parallel. | 
 |     EXPECT_LE(epochMsAfter, epochMsBefore + 3 * sleepMs); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, ThreadPoolOverSaturated) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     constexpr size_t kNumThreads = 10; | 
 |     constexpr size_t kNumCalls = kNumThreads + 3; | 
 |     auto proc = createRpcTestSocketServerProcess({.numThreads = kNumThreads}); | 
 |     testThreadPoolOverSaturated(proc.rootIface, kNumCalls); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, ThreadPoolLimitOutgoing) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     constexpr size_t kNumThreads = 20; | 
 |     constexpr size_t kNumOutgoingConnections = 10; | 
 |     constexpr size_t kNumCalls = kNumOutgoingConnections + 3; | 
 |     auto proc = createRpcTestSocketServerProcess( | 
 |             {.numThreads = kNumThreads, .numOutgoingConnections = kNumOutgoingConnections}); | 
 |     testThreadPoolOverSaturated(proc.rootIface, kNumCalls); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, ThreadingStressTest) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     constexpr size_t kNumClientThreads = 10; | 
 |     constexpr size_t kNumServerThreads = 10; | 
 |     constexpr size_t kNumCalls = 100; | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({.numThreads = kNumServerThreads}); | 
 |  | 
 |     std::vector<std::thread> threads; | 
 |     for (size_t i = 0; i < kNumClientThreads; i++) { | 
 |         threads.push_back(std::thread([&] { | 
 |             for (size_t j = 0; j < kNumCalls; j++) { | 
 |                 sp<IBinder> out; | 
 |                 EXPECT_OK(proc.rootIface->repeatBinder(proc.rootBinder, &out)); | 
 |                 EXPECT_EQ(proc.rootBinder, out); | 
 |             } | 
 |         })); | 
 |     } | 
 |  | 
 |     for (auto& t : threads) t.join(); | 
 | } | 
 |  | 
 | static void saturateThreadPool(size_t threadCount, const sp<IBinderRpcTest>& iface) { | 
 |     std::vector<std::thread> threads; | 
 |     for (size_t i = 0; i < threadCount; i++) { | 
 |         threads.push_back(std::thread([&] { EXPECT_OK(iface->sleepMs(500)); })); | 
 |     } | 
 |     for (auto& t : threads) t.join(); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, OnewayStressTest) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     constexpr size_t kNumClientThreads = 10; | 
 |     constexpr size_t kNumServerThreads = 10; | 
 |     constexpr size_t kNumCalls = 1000; | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({.numThreads = kNumServerThreads}); | 
 |  | 
 |     std::vector<std::thread> threads; | 
 |     for (size_t i = 0; i < kNumClientThreads; i++) { | 
 |         threads.push_back(std::thread([&] { | 
 |             for (size_t j = 0; j < kNumCalls; j++) { | 
 |                 EXPECT_OK(proc.rootIface->sendString("a")); | 
 |             } | 
 |         })); | 
 |     } | 
 |  | 
 |     for (auto& t : threads) t.join(); | 
 |  | 
 |     saturateThreadPool(kNumServerThreads, proc.rootIface); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, OnewayCallDoesNotWait) { | 
 |     constexpr size_t kReallyLongTimeMs = 100; | 
 |     constexpr size_t kSleepMs = kReallyLongTimeMs * 5; | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     size_t epochMsBefore = epochMillis(); | 
 |  | 
 |     EXPECT_OK(proc.rootIface->sleepMsAsync(kSleepMs)); | 
 |  | 
 |     size_t epochMsAfter = epochMillis(); | 
 |     EXPECT_LT(epochMsAfter, epochMsBefore + kReallyLongTimeMs); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, OnewayCallQueueingWithFds) { | 
 |     if (!supportsFdTransport()) { | 
 |         GTEST_SKIP() << "Would fail trivially (which is tested elsewhere)"; | 
 |     } | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     // This test forces a oneway transaction to be queued by issuing two | 
 |     // `blockingSendFdOneway` calls, then drains the queue by issuing two | 
 |     // `blockingRecvFd` calls. | 
 |     // | 
 |     // For more details about the queuing semantics see | 
 |     // https://developer.android.com/reference/android/os/IBinder#FLAG_ONEWAY | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             .numThreads = 3, | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::UNIX, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::UNIX}, | 
 |     }); | 
 |  | 
 |     EXPECT_OK(proc.rootIface->blockingSendFdOneway( | 
 |             android::os::ParcelFileDescriptor(mockFileDescriptor("a")))); | 
 |     EXPECT_OK(proc.rootIface->blockingSendFdOneway( | 
 |             android::os::ParcelFileDescriptor(mockFileDescriptor("b")))); | 
 |  | 
 |     android::os::ParcelFileDescriptor fdA; | 
 |     EXPECT_OK(proc.rootIface->blockingRecvFd(&fdA)); | 
 |     std::string result; | 
 |     CHECK(android::base::ReadFdToString(fdA.get(), &result)); | 
 |     EXPECT_EQ(result, "a"); | 
 |  | 
 |     android::os::ParcelFileDescriptor fdB; | 
 |     EXPECT_OK(proc.rootIface->blockingRecvFd(&fdB)); | 
 |     CHECK(android::base::ReadFdToString(fdB.get(), &result)); | 
 |     EXPECT_EQ(result, "b"); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, OnewayCallQueueing) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     constexpr size_t kNumSleeps = 10; | 
 |     constexpr size_t kNumExtraServerThreads = 4; | 
 |     constexpr size_t kSleepMs = 50; | 
 |  | 
 |     // make sure calls to the same object happen on the same thread | 
 |     auto proc = createRpcTestSocketServerProcess({.numThreads = 1 + kNumExtraServerThreads}); | 
 |  | 
 |     EXPECT_OK(proc.rootIface->lock()); | 
 |  | 
 |     size_t epochMsBefore = epochMillis(); | 
 |  | 
 |     // all these *Async commands should be queued on the server sequentially, | 
 |     // even though there are multiple threads. | 
 |     for (size_t i = 0; i + 1 < kNumSleeps; i++) { | 
 |         proc.rootIface->sleepMsAsync(kSleepMs); | 
 |     } | 
 |     EXPECT_OK(proc.rootIface->unlockInMsAsync(kSleepMs)); | 
 |  | 
 |     // this can only return once the final async call has unlocked | 
 |     EXPECT_OK(proc.rootIface->lockUnlock()); | 
 |  | 
 |     size_t epochMsAfter = epochMillis(); | 
 |  | 
 |     EXPECT_GE(epochMsAfter, epochMsBefore + kSleepMs * kNumSleeps); | 
 |  | 
 |     saturateThreadPool(1 + kNumExtraServerThreads, proc.rootIface); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, OnewayCallExhaustion) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     constexpr size_t kNumClients = 2; | 
 |     constexpr size_t kTooLongMs = 1000; | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({.numThreads = kNumClients, .numSessions = 2}); | 
 |  | 
 |     // Build up oneway calls on the second session to make sure it terminates | 
 |     // and shuts down. The first session should be unaffected (proc destructor | 
 |     // checks the first session). | 
 |     auto iface = interface_cast<IBinderRpcTest>(proc.proc.sessions.at(1).root); | 
 |  | 
 |     std::vector<std::thread> threads; | 
 |     for (size_t i = 0; i < kNumClients; i++) { | 
 |         // one of these threads will get stuck queueing a transaction once the | 
 |         // socket fills up, the other will be able to fill up transactions on | 
 |         // this object | 
 |         threads.push_back(std::thread([&] { | 
 |             while (iface->sleepMsAsync(kTooLongMs).isOk()) { | 
 |             } | 
 |         })); | 
 |     } | 
 |     for (auto& t : threads) t.join(); | 
 |  | 
 |     Status status = iface->sleepMsAsync(kTooLongMs); | 
 |     EXPECT_EQ(DEAD_OBJECT, status.transactionError()) << status; | 
 |  | 
 |     // now that it has died, wait for the remote session to shutdown | 
 |     std::vector<int32_t> remoteCounts; | 
 |     do { | 
 |         EXPECT_OK(proc.rootIface->countBinders(&remoteCounts)); | 
 |     } while (remoteCounts.size() == kNumClients); | 
 |  | 
 |     // the second session should be shutdown in the other process by the time we | 
 |     // are able to join above (it'll only be hung up once it finishes processing | 
 |     // any pending commands). We need to erase this session from the record | 
 |     // here, so that the destructor for our session won't check that this | 
 |     // session is valid, but we still want it to test the other session. | 
 |     proc.proc.sessions.erase(proc.proc.sessions.begin() + 1); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, Callbacks) { | 
 |     const static std::string kTestString = "good afternoon!"; | 
 |  | 
 |     for (bool callIsOneway : {true, false}) { | 
 |         for (bool callbackIsOneway : {true, false}) { | 
 |             for (bool delayed : {true, false}) { | 
 |                 if (clientOrServerSingleThreaded() && | 
 |                     (callIsOneway || callbackIsOneway || delayed)) { | 
 |                     // we have no incoming connections to receive the callback | 
 |                     continue; | 
 |                 } | 
 |  | 
 |                 size_t numIncomingConnections = clientOrServerSingleThreaded() ? 0 : 1; | 
 |                 auto proc = createRpcTestSocketServerProcess( | 
 |                         {.numThreads = 1, | 
 |                          .numSessions = 1, | 
 |                          .numIncomingConnections = numIncomingConnections}); | 
 |                 auto cb = sp<MyBinderRpcCallback>::make(); | 
 |  | 
 |                 if (callIsOneway) { | 
 |                     EXPECT_OK(proc.rootIface->doCallbackAsync(cb, callbackIsOneway, delayed, | 
 |                                                               kTestString)); | 
 |                 } else { | 
 |                     EXPECT_OK( | 
 |                             proc.rootIface->doCallback(cb, callbackIsOneway, delayed, kTestString)); | 
 |                 } | 
 |  | 
 |                 // if both transactions are synchronous and the response is sent back on the | 
 |                 // same thread, everything should have happened in a nested call. Otherwise, | 
 |                 // the callback will be processed on another thread. | 
 |                 if (callIsOneway || callbackIsOneway || delayed) { | 
 |                     using std::literals::chrono_literals::operator""s; | 
 |                     RpcMutexUniqueLock _l(cb->mMutex); | 
 |                     cb->mCv.wait_for(_l, 1s, [&] { return !cb->mValues.empty(); }); | 
 |                 } | 
 |  | 
 |                 EXPECT_EQ(cb->mValues.size(), 1) | 
 |                         << "callIsOneway: " << callIsOneway | 
 |                         << " callbackIsOneway: " << callbackIsOneway << " delayed: " << delayed; | 
 |                 if (cb->mValues.empty()) continue; | 
 |                 EXPECT_EQ(cb->mValues.at(0), kTestString) | 
 |                         << "callIsOneway: " << callIsOneway | 
 |                         << " callbackIsOneway: " << callbackIsOneway << " delayed: " << delayed; | 
 |  | 
 |                 // since we are severing the connection, we need to go ahead and | 
 |                 // tell the server to shutdown and exit so that waitpid won't hang | 
 |                 if (auto status = proc.rootIface->scheduleShutdown(); !status.isOk()) { | 
 |                     EXPECT_EQ(DEAD_OBJECT, status.transactionError()) << status; | 
 |                 } | 
 |  | 
 |                 // since this session has an incoming connection w/ a threadpool, we | 
 |                 // need to manually shut it down | 
 |                 EXPECT_TRUE(proc.proc.sessions.at(0).session->shutdownAndWait(true)); | 
 |                 proc.expectAlreadyShutdown = true; | 
 |             } | 
 |         } | 
 |     } | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SingleDeathRecipient) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |     class MyDeathRec : public IBinder::DeathRecipient { | 
 |     public: | 
 |         void binderDied(const wp<IBinder>& /* who */) override { | 
 |             dead = true; | 
 |             mCv.notify_one(); | 
 |         } | 
 |         std::mutex mMtx; | 
 |         std::condition_variable mCv; | 
 |         bool dead = false; | 
 |     }; | 
 |  | 
 |     // Death recipient needs to have an incoming connection to be called | 
 |     auto proc = createRpcTestSocketServerProcess( | 
 |             {.numThreads = 1, .numSessions = 1, .numIncomingConnections = 1}); | 
 |  | 
 |     auto dr = sp<MyDeathRec>::make(); | 
 |     ASSERT_EQ(OK, proc.rootBinder->linkToDeath(dr, (void*)1, 0)); | 
 |  | 
 |     if (auto status = proc.rootIface->scheduleShutdown(); !status.isOk()) { | 
 |         EXPECT_EQ(DEAD_OBJECT, status.transactionError()) << status; | 
 |     } | 
 |  | 
 |     std::unique_lock<std::mutex> lock(dr->mMtx); | 
 |     ASSERT_TRUE(dr->mCv.wait_for(lock, 100ms, [&]() { return dr->dead; })); | 
 |  | 
 |     // need to wait for the session to shutdown so we don't "Leak session" | 
 |     EXPECT_TRUE(proc.proc.sessions.at(0).session->shutdownAndWait(true)); | 
 |     proc.expectAlreadyShutdown = true; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SingleDeathRecipientOnShutdown) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |     class MyDeathRec : public IBinder::DeathRecipient { | 
 |     public: | 
 |         void binderDied(const wp<IBinder>& /* who */) override { | 
 |             dead = true; | 
 |             mCv.notify_one(); | 
 |         } | 
 |         std::mutex mMtx; | 
 |         std::condition_variable mCv; | 
 |         bool dead = false; | 
 |     }; | 
 |  | 
 |     // Death recipient needs to have an incoming connection to be called | 
 |     auto proc = createRpcTestSocketServerProcess( | 
 |             {.numThreads = 1, .numSessions = 1, .numIncomingConnections = 1}); | 
 |  | 
 |     auto dr = sp<MyDeathRec>::make(); | 
 |     EXPECT_EQ(OK, proc.rootBinder->linkToDeath(dr, (void*)1, 0)); | 
 |  | 
 |     // Explicitly calling shutDownAndWait will cause the death recipients | 
 |     // to be called. | 
 |     EXPECT_TRUE(proc.proc.sessions.at(0).session->shutdownAndWait(true)); | 
 |  | 
 |     std::unique_lock<std::mutex> lock(dr->mMtx); | 
 |     if (!dr->dead) { | 
 |         EXPECT_EQ(std::cv_status::no_timeout, dr->mCv.wait_for(lock, 100ms)); | 
 |     } | 
 |     EXPECT_TRUE(dr->dead) << "Failed to receive the death notification."; | 
 |  | 
 |     proc.proc.host.terminate(); | 
 |     proc.proc.host.setCustomExitStatusCheck([](int wstatus) { | 
 |         EXPECT_TRUE(WIFSIGNALED(wstatus) && WTERMSIG(wstatus) == SIGTERM) | 
 |                 << "server process failed incorrectly: " << WaitStatusToString(wstatus); | 
 |     }); | 
 |     proc.expectAlreadyShutdown = true; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, DeathRecipientFatalWithoutIncoming) { | 
 |     class MyDeathRec : public IBinder::DeathRecipient { | 
 |     public: | 
 |         void binderDied(const wp<IBinder>& /* who */) override {} | 
 |     }; | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess( | 
 |             {.numThreads = 1, .numSessions = 1, .numIncomingConnections = 0}); | 
 |  | 
 |     auto dr = sp<MyDeathRec>::make(); | 
 |     EXPECT_DEATH(proc.rootBinder->linkToDeath(dr, (void*)1, 0), | 
 |                  "Cannot register a DeathRecipient without any incoming connections."); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, UnlinkDeathRecipient) { | 
 |     if (clientOrServerSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |     class MyDeathRec : public IBinder::DeathRecipient { | 
 |     public: | 
 |         void binderDied(const wp<IBinder>& /* who */) override { | 
 |             GTEST_FAIL() << "This should not be called after unlinkToDeath"; | 
 |         } | 
 |     }; | 
 |  | 
 |     // Death recipient needs to have an incoming connection to be called | 
 |     auto proc = createRpcTestSocketServerProcess( | 
 |             {.numThreads = 1, .numSessions = 1, .numIncomingConnections = 1}); | 
 |  | 
 |     auto dr = sp<MyDeathRec>::make(); | 
 |     ASSERT_EQ(OK, proc.rootBinder->linkToDeath(dr, (void*)1, 0)); | 
 |     ASSERT_EQ(OK, proc.rootBinder->unlinkToDeath(dr, (void*)1, 0, nullptr)); | 
 |  | 
 |     if (auto status = proc.rootIface->scheduleShutdown(); !status.isOk()) { | 
 |         EXPECT_EQ(DEAD_OBJECT, status.transactionError()) << status; | 
 |     } | 
 |  | 
 |     // need to wait for the session to shutdown so we don't "Leak session" | 
 |     EXPECT_TRUE(proc.proc.sessions.at(0).session->shutdownAndWait(true)); | 
 |     proc.expectAlreadyShutdown = true; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, OnewayCallbackWithNoThread) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     auto cb = sp<MyBinderRpcCallback>::make(); | 
 |  | 
 |     Status status = proc.rootIface->doCallback(cb, true /*oneway*/, false /*delayed*/, "anything"); | 
 |     EXPECT_EQ(WOULD_BLOCK, status.transactionError()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, Die) { | 
 |     for (bool doDeathCleanup : {true, false}) { | 
 |         auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |         // make sure there is some state during crash | 
 |         // 1. we hold their binder | 
 |         sp<IBinderRpcSession> session; | 
 |         EXPECT_OK(proc.rootIface->openSession("happy", &session)); | 
 |         // 2. they hold our binder | 
 |         sp<IBinder> binder = new BBinder(); | 
 |         EXPECT_OK(proc.rootIface->holdBinder(binder)); | 
 |  | 
 |         EXPECT_EQ(DEAD_OBJECT, proc.rootIface->die(doDeathCleanup).transactionError()) | 
 |                 << "Do death cleanup: " << doDeathCleanup; | 
 |  | 
 |         proc.proc.host.setCustomExitStatusCheck([](int wstatus) { | 
 |             EXPECT_TRUE(WIFEXITED(wstatus) && WEXITSTATUS(wstatus) == 1) | 
 |                     << "server process failed incorrectly: " << WaitStatusToString(wstatus); | 
 |         }); | 
 |         proc.expectAlreadyShutdown = true; | 
 |     } | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, UseKernelBinderCallingId) { | 
 |     // This test only works if the current process shared the internal state of | 
 |     // ProcessState with the service across the call to fork(). Both the static | 
 |     // libraries and libbinder.so have their own separate copies of all the | 
 |     // globals, so the test only works when the test client and service both use | 
 |     // libbinder.so (when using static libraries, even a client and service | 
 |     // using the same kind of static library should have separate copies of the | 
 |     // variables). | 
 |     if (!kEnableSharedLibs || serverSingleThreaded() || noKernel()) { | 
 |         GTEST_SKIP() << "Test disabled because Binder kernel driver was disabled " | 
 |                         "at build time."; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     // we can't allocate IPCThreadState so actually the first time should | 
 |     // succeed :( | 
 |     EXPECT_OK(proc.rootIface->useKernelBinderCallingId()); | 
 |  | 
 |     // second time! we catch the error :) | 
 |     EXPECT_EQ(DEAD_OBJECT, proc.rootIface->useKernelBinderCallingId().transactionError()); | 
 |  | 
 |     proc.proc.host.setCustomExitStatusCheck([](int wstatus) { | 
 |         EXPECT_TRUE(WIFSIGNALED(wstatus) && WTERMSIG(wstatus) == SIGABRT) | 
 |                 << "server process failed incorrectly: " << WaitStatusToString(wstatus); | 
 |     }); | 
 |     proc.expectAlreadyShutdown = true; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, FileDescriptorTransportRejectNone) { | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::NONE, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::UNIX}, | 
 |             .allowConnectFailure = true, | 
 |     }); | 
 |     EXPECT_TRUE(proc.proc.sessions.empty()) << "session connections should have failed"; | 
 |     proc.proc.host.terminate(); | 
 |     proc.proc.host.setCustomExitStatusCheck([](int wstatus) { | 
 |         EXPECT_TRUE(WIFSIGNALED(wstatus) && WTERMSIG(wstatus) == SIGTERM) | 
 |                 << "server process failed incorrectly: " << WaitStatusToString(wstatus); | 
 |     }); | 
 |     proc.expectAlreadyShutdown = true; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, FileDescriptorTransportRejectUnix) { | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::UNIX, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::NONE}, | 
 |             .allowConnectFailure = true, | 
 |     }); | 
 |     EXPECT_TRUE(proc.proc.sessions.empty()) << "session connections should have failed"; | 
 |     proc.proc.host.terminate(); | 
 |     proc.proc.host.setCustomExitStatusCheck([](int wstatus) { | 
 |         EXPECT_TRUE(WIFSIGNALED(wstatus) && WTERMSIG(wstatus) == SIGTERM) | 
 |                 << "server process failed incorrectly: " << WaitStatusToString(wstatus); | 
 |     }); | 
 |     proc.expectAlreadyShutdown = true; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, FileDescriptorTransportOptionalUnix) { | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::NONE, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::NONE, | 
 |                      RpcSession::FileDescriptorTransportMode::UNIX}, | 
 |     }); | 
 |  | 
 |     android::os::ParcelFileDescriptor out; | 
 |     auto status = proc.rootIface->echoAsFile("hello", &out); | 
 |     EXPECT_EQ(status.transactionError(), FDS_NOT_ALLOWED) << status; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, ReceiveFile) { | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::UNIX, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::UNIX}, | 
 |     }); | 
 |  | 
 |     android::os::ParcelFileDescriptor out; | 
 |     auto status = proc.rootIface->echoAsFile("hello", &out); | 
 |     if (!supportsFdTransport()) { | 
 |         EXPECT_EQ(status.transactionError(), BAD_VALUE) << status; | 
 |         return; | 
 |     } | 
 |     ASSERT_TRUE(status.isOk()) << status; | 
 |  | 
 |     std::string result; | 
 |     CHECK(android::base::ReadFdToString(out.get(), &result)); | 
 |     EXPECT_EQ(result, "hello"); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SendFiles) { | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::UNIX, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::UNIX}, | 
 |     }); | 
 |  | 
 |     std::vector<android::os::ParcelFileDescriptor> files; | 
 |     files.emplace_back(android::os::ParcelFileDescriptor(mockFileDescriptor("123"))); | 
 |     files.emplace_back(android::os::ParcelFileDescriptor(mockFileDescriptor("a"))); | 
 |     files.emplace_back(android::os::ParcelFileDescriptor(mockFileDescriptor("b"))); | 
 |     files.emplace_back(android::os::ParcelFileDescriptor(mockFileDescriptor("cd"))); | 
 |  | 
 |     android::os::ParcelFileDescriptor out; | 
 |     auto status = proc.rootIface->concatFiles(files, &out); | 
 |     if (!supportsFdTransport()) { | 
 |         EXPECT_EQ(status.transactionError(), BAD_VALUE) << status; | 
 |         return; | 
 |     } | 
 |     ASSERT_TRUE(status.isOk()) << status; | 
 |  | 
 |     std::string result; | 
 |     CHECK(android::base::ReadFdToString(out.get(), &result)); | 
 |     EXPECT_EQ(result, "123abcd"); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SendMaxFiles) { | 
 |     if (!supportsFdTransport()) { | 
 |         GTEST_SKIP() << "Would fail trivially (which is tested by BinderRpc::SendFiles)"; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::UNIX, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::UNIX}, | 
 |     }); | 
 |  | 
 |     std::vector<android::os::ParcelFileDescriptor> files; | 
 |     for (int i = 0; i < 253; i++) { | 
 |         files.emplace_back(android::os::ParcelFileDescriptor(mockFileDescriptor("a"))); | 
 |     } | 
 |  | 
 |     android::os::ParcelFileDescriptor out; | 
 |     auto status = proc.rootIface->concatFiles(files, &out); | 
 |     ASSERT_TRUE(status.isOk()) << status; | 
 |  | 
 |     std::string result; | 
 |     CHECK(android::base::ReadFdToString(out.get(), &result)); | 
 |     EXPECT_EQ(result, std::string(253, 'a')); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, SendTooManyFiles) { | 
 |     if (!supportsFdTransport()) { | 
 |         GTEST_SKIP() << "Would fail trivially (which is tested by BinderRpc::SendFiles)"; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({ | 
 |             .clientFileDescriptorTransportMode = RpcSession::FileDescriptorTransportMode::UNIX, | 
 |             .serverSupportedFileDescriptorTransportModes = | 
 |                     {RpcSession::FileDescriptorTransportMode::UNIX}, | 
 |     }); | 
 |  | 
 |     std::vector<android::os::ParcelFileDescriptor> files; | 
 |     for (int i = 0; i < 254; i++) { | 
 |         files.emplace_back(android::os::ParcelFileDescriptor(mockFileDescriptor("a"))); | 
 |     } | 
 |  | 
 |     android::os::ParcelFileDescriptor out; | 
 |     auto status = proc.rootIface->concatFiles(files, &out); | 
 |     EXPECT_EQ(status.transactionError(), BAD_VALUE) << status; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, WorksWithLibbinderNdkPing) { | 
 |     if constexpr (!kEnableSharedLibs) { | 
 |         GTEST_SKIP() << "Test disabled because Binder was built as a static library"; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     ndk::SpAIBinder binder = ndk::SpAIBinder(AIBinder_fromPlatformBinder(proc.rootBinder)); | 
 |     ASSERT_NE(binder, nullptr); | 
 |  | 
 |     ASSERT_EQ(STATUS_OK, AIBinder_ping(binder.get())); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, WorksWithLibbinderNdkUserTransaction) { | 
 |     if constexpr (!kEnableSharedLibs) { | 
 |         GTEST_SKIP() << "Test disabled because Binder was built as a static library"; | 
 |     } | 
 |  | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |  | 
 |     ndk::SpAIBinder binder = ndk::SpAIBinder(AIBinder_fromPlatformBinder(proc.rootBinder)); | 
 |     ASSERT_NE(binder, nullptr); | 
 |  | 
 |     auto ndkBinder = aidl::IBinderRpcTest::fromBinder(binder); | 
 |     ASSERT_NE(ndkBinder, nullptr); | 
 |  | 
 |     std::string out; | 
 |     ndk::ScopedAStatus status = ndkBinder->doubleString("aoeu", &out); | 
 |     ASSERT_TRUE(status.isOk()) << status.getDescription(); | 
 |     ASSERT_EQ("aoeuaoeu", out); | 
 | } | 
 |  | 
 | ssize_t countFds() { | 
 |     DIR* dir = opendir("/proc/self/fd/"); | 
 |     if (dir == nullptr) return -1; | 
 |     ssize_t ret = 0; | 
 |     dirent* ent; | 
 |     while ((ent = readdir(dir)) != nullptr) ret++; | 
 |     closedir(dir); | 
 |     return ret; | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, Fds) { | 
 |     if (serverSingleThreaded()) { | 
 |         GTEST_SKIP() << "This test requires multiple threads"; | 
 |     } | 
 |  | 
 |     ssize_t beforeFds = countFds(); | 
 |     ASSERT_GE(beforeFds, 0); | 
 |     { | 
 |         auto proc = createRpcTestSocketServerProcess({.numThreads = 10}); | 
 |         ASSERT_EQ(OK, proc.rootBinder->pingBinder()); | 
 |     } | 
 |     ASSERT_EQ(beforeFds, countFds()) << (system("ls -l /proc/self/fd/"), "fd leak?"); | 
 | } | 
 |  | 
 | TEST_P(BinderRpc, AidlDelegatorTest) { | 
 |     auto proc = createRpcTestSocketServerProcess({}); | 
 |     auto myDelegator = sp<IBinderRpcTestDelegator>::make(proc.rootIface); | 
 |     ASSERT_NE(nullptr, myDelegator); | 
 |  | 
 |     std::string doubled; | 
 |     EXPECT_OK(myDelegator->doubleString("cool ", &doubled)); | 
 |     EXPECT_EQ("cool cool ", doubled); | 
 | } | 
 |  | 
 | static bool testSupportVsockLoopback() { | 
 |     // We don't need to enable TLS to know if vsock is supported. | 
 |     unsigned int vsockPort = allocateVsockPort(); | 
 |  | 
 |     android::base::unique_fd serverFd( | 
 |             TEMP_FAILURE_RETRY(socket(AF_VSOCK, SOCK_STREAM | SOCK_CLOEXEC | SOCK_NONBLOCK, 0))); | 
 |     LOG_ALWAYS_FATAL_IF(serverFd == -1, "Could not create socket: %s", strerror(errno)); | 
 |  | 
 |     sockaddr_vm serverAddr{ | 
 |             .svm_family = AF_VSOCK, | 
 |             .svm_port = vsockPort, | 
 |             .svm_cid = VMADDR_CID_ANY, | 
 |     }; | 
 |     int ret = TEMP_FAILURE_RETRY( | 
 |             bind(serverFd.get(), reinterpret_cast<sockaddr*>(&serverAddr), sizeof(serverAddr))); | 
 |     LOG_ALWAYS_FATAL_IF(0 != ret, "Could not bind socket to port %u: %s", vsockPort, | 
 |                         strerror(errno)); | 
 |  | 
 |     ret = TEMP_FAILURE_RETRY(listen(serverFd.get(), 1 /*backlog*/)); | 
 |     LOG_ALWAYS_FATAL_IF(0 != ret, "Could not listen socket on port %u: %s", vsockPort, | 
 |                         strerror(errno)); | 
 |  | 
 |     // Try to connect to the server using the VMADDR_CID_LOCAL cid | 
 |     // to see if the kernel supports it. It's safe to use a blocking | 
 |     // connect because vsock sockets have a 2 second connection timeout, | 
 |     // and they return ETIMEDOUT after that. | 
 |     android::base::unique_fd connectFd( | 
 |             TEMP_FAILURE_RETRY(socket(AF_VSOCK, SOCK_STREAM | SOCK_CLOEXEC | SOCK_NONBLOCK, 0))); | 
 |     LOG_ALWAYS_FATAL_IF(connectFd == -1, "Could not create socket for port %u: %s", vsockPort, | 
 |                         strerror(errno)); | 
 |  | 
 |     bool success = false; | 
 |     sockaddr_vm connectAddr{ | 
 |             .svm_family = AF_VSOCK, | 
 |             .svm_port = vsockPort, | 
 |             .svm_cid = VMADDR_CID_LOCAL, | 
 |     }; | 
 |     ret = TEMP_FAILURE_RETRY(connect(connectFd.get(), reinterpret_cast<sockaddr*>(&connectAddr), | 
 |                                      sizeof(connectAddr))); | 
 |     if (ret != 0 && (errno == EAGAIN || errno == EINPROGRESS)) { | 
 |         android::base::unique_fd acceptFd; | 
 |         while (true) { | 
 |             pollfd pfd[]{ | 
 |                     {.fd = serverFd.get(), .events = POLLIN, .revents = 0}, | 
 |                     {.fd = connectFd.get(), .events = POLLOUT, .revents = 0}, | 
 |             }; | 
 |             ret = TEMP_FAILURE_RETRY(poll(pfd, arraysize(pfd), -1)); | 
 |             LOG_ALWAYS_FATAL_IF(ret < 0, "Error polling: %s", strerror(errno)); | 
 |  | 
 |             if (pfd[0].revents & POLLIN) { | 
 |                 sockaddr_vm acceptAddr; | 
 |                 socklen_t acceptAddrLen = sizeof(acceptAddr); | 
 |                 ret = TEMP_FAILURE_RETRY(accept4(serverFd.get(), | 
 |                                                  reinterpret_cast<sockaddr*>(&acceptAddr), | 
 |                                                  &acceptAddrLen, SOCK_CLOEXEC)); | 
 |                 LOG_ALWAYS_FATAL_IF(ret < 0, "Could not accept4 socket: %s", strerror(errno)); | 
 |                 LOG_ALWAYS_FATAL_IF(acceptAddrLen != static_cast<socklen_t>(sizeof(acceptAddr)), | 
 |                                     "Truncated address"); | 
 |  | 
 |                 // Store the fd in acceptFd so we keep the connection alive | 
 |                 // while polling connectFd | 
 |                 acceptFd.reset(ret); | 
 |             } | 
 |  | 
 |             if (pfd[1].revents & POLLOUT) { | 
 |                 // Connect either succeeded or timed out | 
 |                 int connectErrno; | 
 |                 socklen_t connectErrnoLen = sizeof(connectErrno); | 
 |                 int ret = getsockopt(connectFd.get(), SOL_SOCKET, SO_ERROR, &connectErrno, | 
 |                                      &connectErrnoLen); | 
 |                 LOG_ALWAYS_FATAL_IF(ret == -1, | 
 |                                     "Could not getsockopt() after connect() " | 
 |                                     "on non-blocking socket: %s.", | 
 |                                     strerror(errno)); | 
 |  | 
 |                 // We're done, this is all we wanted | 
 |                 success = connectErrno == 0; | 
 |                 break; | 
 |             } | 
 |         } | 
 |     } else { | 
 |         success = ret == 0; | 
 |     } | 
 |  | 
 |     ALOGE("Detected vsock loopback supported: %s", success ? "yes" : "no"); | 
 |  | 
 |     return success; | 
 | } | 
 |  | 
 | static std::vector<SocketType> testSocketTypes(bool hasPreconnected = true) { | 
 |     std::vector<SocketType> ret = {SocketType::UNIX, SocketType::UNIX_BOOTSTRAP, SocketType::INET}; | 
 |  | 
 |     if (hasPreconnected) ret.push_back(SocketType::PRECONNECTED); | 
 |  | 
 |     static bool hasVsockLoopback = testSupportVsockLoopback(); | 
 |  | 
 |     if (hasVsockLoopback) { | 
 |         ret.push_back(SocketType::VSOCK); | 
 |     } | 
 |  | 
 |     return ret; | 
 | } | 
 |  | 
 | static std::vector<uint32_t> testVersions() { | 
 |     std::vector<uint32_t> versions; | 
 |     for (size_t i = 0; i < RPC_WIRE_PROTOCOL_VERSION_NEXT; i++) { | 
 |         versions.push_back(i); | 
 |     } | 
 |     versions.push_back(RPC_WIRE_PROTOCOL_VERSION_EXPERIMENTAL); | 
 |     return versions; | 
 | } | 
 |  | 
 | INSTANTIATE_TEST_CASE_P(PerSocket, BinderRpc, | 
 |                         ::testing::Combine(::testing::ValuesIn(testSocketTypes()), | 
 |                                            ::testing::ValuesIn(RpcSecurityValues()), | 
 |                                            ::testing::ValuesIn(testVersions()), | 
 |                                            ::testing::ValuesIn(testVersions()), | 
 |                                            ::testing::Values(false, true), | 
 |                                            ::testing::Values(false, true)), | 
 |                         BinderRpc::PrintParamInfo); | 
 |  | 
 | class BinderRpcServerRootObject | 
 |       : public ::testing::TestWithParam<std::tuple<bool, bool, RpcSecurity>> {}; | 
 |  | 
 | TEST_P(BinderRpcServerRootObject, WeakRootObject) { | 
 |     using SetFn = std::function<void(RpcServer*, sp<IBinder>)>; | 
 |     auto setRootObject = [](bool isStrong) -> SetFn { | 
 |         return isStrong ? SetFn(&RpcServer::setRootObject) : SetFn(&RpcServer::setRootObjectWeak); | 
 |     }; | 
 |  | 
 |     auto [isStrong1, isStrong2, rpcSecurity] = GetParam(); | 
 |     auto server = RpcServer::make(newFactory(rpcSecurity)); | 
 |     auto binder1 = sp<BBinder>::make(); | 
 |     IBinder* binderRaw1 = binder1.get(); | 
 |     setRootObject(isStrong1)(server.get(), binder1); | 
 |     EXPECT_EQ(binderRaw1, server->getRootObject()); | 
 |     binder1.clear(); | 
 |     EXPECT_EQ((isStrong1 ? binderRaw1 : nullptr), server->getRootObject()); | 
 |  | 
 |     auto binder2 = sp<BBinder>::make(); | 
 |     IBinder* binderRaw2 = binder2.get(); | 
 |     setRootObject(isStrong2)(server.get(), binder2); | 
 |     EXPECT_EQ(binderRaw2, server->getRootObject()); | 
 |     binder2.clear(); | 
 |     EXPECT_EQ((isStrong2 ? binderRaw2 : nullptr), server->getRootObject()); | 
 | } | 
 |  | 
 | INSTANTIATE_TEST_CASE_P(BinderRpc, BinderRpcServerRootObject, | 
 |                         ::testing::Combine(::testing::Bool(), ::testing::Bool(), | 
 |                                            ::testing::ValuesIn(RpcSecurityValues()))); | 
 |  | 
 | class OneOffSignal { | 
 | public: | 
 |     // If notify() was previously called, or is called within |duration|, return true; else false. | 
 |     template <typename R, typename P> | 
 |     bool wait(std::chrono::duration<R, P> duration) { | 
 |         std::unique_lock<std::mutex> lock(mMutex); | 
 |         return mCv.wait_for(lock, duration, [this] { return mValue; }); | 
 |     } | 
 |     void notify() { | 
 |         std::unique_lock<std::mutex> lock(mMutex); | 
 |         mValue = true; | 
 |         lock.unlock(); | 
 |         mCv.notify_all(); | 
 |     } | 
 |  | 
 | private: | 
 |     std::mutex mMutex; | 
 |     std::condition_variable mCv; | 
 |     bool mValue = false; | 
 | }; | 
 |  | 
 | TEST(BinderRpc, Java) { | 
 | #if !defined(__ANDROID__) | 
 |     GTEST_SKIP() << "This test is only run on Android. Though it can technically run on host on" | 
 |                     "createRpcDelegateServiceManager() with a device attached, such test belongs " | 
 |                     "to binderHostDeviceTest. Hence, just disable this test on host."; | 
 | #endif // !__ANDROID__ | 
 |     if constexpr (!kEnableKernelIpc) { | 
 |         GTEST_SKIP() << "Test disabled because Binder kernel driver was disabled " | 
 |                         "at build time."; | 
 |     } | 
 |  | 
 |     sp<IServiceManager> sm = defaultServiceManager(); | 
 |     ASSERT_NE(nullptr, sm); | 
 |     // Any Java service with non-empty getInterfaceDescriptor() would do. | 
 |     // Let's pick batteryproperties. | 
 |     auto binder = sm->checkService(String16("batteryproperties")); | 
 |     ASSERT_NE(nullptr, binder); | 
 |     auto descriptor = binder->getInterfaceDescriptor(); | 
 |     ASSERT_GE(descriptor.size(), 0); | 
 |     ASSERT_EQ(OK, binder->pingBinder()); | 
 |  | 
 |     auto rpcServer = RpcServer::make(); | 
 |     unsigned int port; | 
 |     ASSERT_EQ(OK, rpcServer->setupInetServer(kLocalInetAddress, 0, &port)); | 
 |     auto socket = rpcServer->releaseServer(); | 
 |  | 
 |     auto keepAlive = sp<BBinder>::make(); | 
 |     auto setRpcClientDebugStatus = binder->setRpcClientDebug(std::move(socket), keepAlive); | 
 |  | 
 |     if (!android::base::GetBoolProperty("ro.debuggable", false) || | 
 |         android::base::GetProperty("ro.build.type", "") == "user") { | 
 |         ASSERT_EQ(INVALID_OPERATION, setRpcClientDebugStatus) | 
 |                 << "setRpcClientDebug should return INVALID_OPERATION on non-debuggable or user " | 
 |                    "builds, but get " | 
 |                 << statusToString(setRpcClientDebugStatus); | 
 |         GTEST_SKIP(); | 
 |     } | 
 |  | 
 |     ASSERT_EQ(OK, setRpcClientDebugStatus); | 
 |  | 
 |     auto rpcSession = RpcSession::make(); | 
 |     ASSERT_EQ(OK, rpcSession->setupInetClient("127.0.0.1", port)); | 
 |     auto rpcBinder = rpcSession->getRootObject(); | 
 |     ASSERT_NE(nullptr, rpcBinder); | 
 |  | 
 |     ASSERT_EQ(OK, rpcBinder->pingBinder()); | 
 |  | 
 |     ASSERT_EQ(descriptor, rpcBinder->getInterfaceDescriptor()) | 
 |             << "getInterfaceDescriptor should not crash system_server"; | 
 |     ASSERT_EQ(OK, rpcBinder->pingBinder()); | 
 | } | 
 |  | 
 | class BinderRpcServerOnly : public ::testing::TestWithParam<std::tuple<RpcSecurity, uint32_t>> { | 
 | public: | 
 |     static std::string PrintTestParam(const ::testing::TestParamInfo<ParamType>& info) { | 
 |         return std::string(newFactory(std::get<0>(info.param))->toCString()) + "_serverV" + | 
 |                 std::to_string(std::get<1>(info.param)); | 
 |     } | 
 | }; | 
 |  | 
 | TEST_P(BinderRpcServerOnly, SetExternalServerTest) { | 
 |     base::unique_fd sink(TEMP_FAILURE_RETRY(open("/dev/null", O_RDWR))); | 
 |     int sinkFd = sink.get(); | 
 |     auto server = RpcServer::make(newFactory(std::get<0>(GetParam()))); | 
 |     server->setProtocolVersion(std::get<1>(GetParam())); | 
 |     ASSERT_FALSE(server->hasServer()); | 
 |     ASSERT_EQ(OK, server->setupExternalServer(std::move(sink))); | 
 |     ASSERT_TRUE(server->hasServer()); | 
 |     base::unique_fd retrieved = server->releaseServer(); | 
 |     ASSERT_FALSE(server->hasServer()); | 
 |     ASSERT_EQ(sinkFd, retrieved.get()); | 
 | } | 
 |  | 
 | TEST_P(BinderRpcServerOnly, Shutdown) { | 
 |     if constexpr (!kEnableRpcThreads) { | 
 |         GTEST_SKIP() << "Test skipped because threads were disabled at build time"; | 
 |     } | 
 |  | 
 |     auto addr = allocateSocketAddress(); | 
 |     auto server = RpcServer::make(newFactory(std::get<0>(GetParam()))); | 
 |     server->setProtocolVersion(std::get<1>(GetParam())); | 
 |     ASSERT_EQ(OK, server->setupUnixDomainServer(addr.c_str())); | 
 |     auto joinEnds = std::make_shared<OneOffSignal>(); | 
 |  | 
 |     // If things are broken and the thread never stops, don't block other tests. Because the thread | 
 |     // may run after the test finishes, it must not access the stack memory of the test. Hence, | 
 |     // shared pointers are passed. | 
 |     std::thread([server, joinEnds] { | 
 |         server->join(); | 
 |         joinEnds->notify(); | 
 |     }).detach(); | 
 |  | 
 |     bool shutdown = false; | 
 |     for (int i = 0; i < 10 && !shutdown; i++) { | 
 |         usleep(30 * 1000); // 30ms; total 300ms | 
 |         if (server->shutdown()) shutdown = true; | 
 |     } | 
 |     ASSERT_TRUE(shutdown) << "server->shutdown() never returns true"; | 
 |  | 
 |     ASSERT_TRUE(joinEnds->wait(2s)) | 
 |             << "After server->shutdown() returns true, join() did not stop after 2s"; | 
 | } | 
 |  | 
 | INSTANTIATE_TEST_CASE_P(BinderRpc, BinderRpcServerOnly, | 
 |                         ::testing::Combine(::testing::ValuesIn(RpcSecurityValues()), | 
 |                                            ::testing::ValuesIn(testVersions())), | 
 |                         BinderRpcServerOnly::PrintTestParam); | 
 |  | 
 | class RpcTransportTestUtils { | 
 | public: | 
 |     // Only parameterized only server version because `RpcSession` is bypassed | 
 |     // in the client half of the tests. | 
 |     using Param = | 
 |             std::tuple<SocketType, RpcSecurity, std::optional<RpcCertificateFormat>, uint32_t>; | 
 |     using ConnectToServer = std::function<base::unique_fd()>; | 
 |  | 
 |     // A server that handles client socket connections. | 
 |     class Server { | 
 |     public: | 
 |         using AcceptConnection = std::function<base::unique_fd(Server*)>; | 
 |  | 
 |         explicit Server() {} | 
 |         Server(Server&&) = default; | 
 |         ~Server() { shutdownAndWait(); } | 
 |         [[nodiscard]] AssertionResult setUp( | 
 |                 const Param& param, | 
 |                 std::unique_ptr<RpcAuth> auth = std::make_unique<RpcAuthSelfSigned>()) { | 
 |             auto [socketType, rpcSecurity, certificateFormat, serverVersion] = param; | 
 |             auto rpcServer = RpcServer::make(newFactory(rpcSecurity)); | 
 |             rpcServer->setProtocolVersion(serverVersion); | 
 |             switch (socketType) { | 
 |                 case SocketType::PRECONNECTED: { | 
 |                     return AssertionFailure() << "Not supported by this test"; | 
 |                 } break; | 
 |                 case SocketType::UNIX: { | 
 |                     auto addr = allocateSocketAddress(); | 
 |                     auto status = rpcServer->setupUnixDomainServer(addr.c_str()); | 
 |                     if (status != OK) { | 
 |                         return AssertionFailure() | 
 |                                 << "setupUnixDomainServer: " << statusToString(status); | 
 |                     } | 
 |                     mConnectToServer = [addr] { | 
 |                         return connectTo(UnixSocketAddress(addr.c_str())); | 
 |                     }; | 
 |                 } break; | 
 |                 case SocketType::UNIX_BOOTSTRAP: { | 
 |                     base::unique_fd bootstrapFdClient, bootstrapFdServer; | 
 |                     if (!base::Socketpair(SOCK_STREAM, &bootstrapFdClient, &bootstrapFdServer)) { | 
 |                         return AssertionFailure() << "Socketpair() failed"; | 
 |                     } | 
 |                     auto status = rpcServer->setupUnixDomainSocketBootstrapServer( | 
 |                             std::move(bootstrapFdServer)); | 
 |                     if (status != OK) { | 
 |                         return AssertionFailure() << "setupUnixDomainSocketBootstrapServer: " | 
 |                                                   << statusToString(status); | 
 |                     } | 
 |                     mBootstrapSocket = RpcTransportFd(std::move(bootstrapFdClient)); | 
 |                     mAcceptConnection = &Server::recvmsgServerConnection; | 
 |                     mConnectToServer = [this] { return connectToUnixBootstrap(mBootstrapSocket); }; | 
 |                 } break; | 
 |                 case SocketType::VSOCK: { | 
 |                     auto port = allocateVsockPort(); | 
 |                     auto status = rpcServer->setupVsockServer(port); | 
 |                     if (status != OK) { | 
 |                         return AssertionFailure() << "setupVsockServer: " << statusToString(status); | 
 |                     } | 
 |                     mConnectToServer = [port] { | 
 |                         return connectTo(VsockSocketAddress(VMADDR_CID_LOCAL, port)); | 
 |                     }; | 
 |                 } break; | 
 |                 case SocketType::INET: { | 
 |                     unsigned int port; | 
 |                     auto status = rpcServer->setupInetServer(kLocalInetAddress, 0, &port); | 
 |                     if (status != OK) { | 
 |                         return AssertionFailure() << "setupInetServer: " << statusToString(status); | 
 |                     } | 
 |                     mConnectToServer = [port] { | 
 |                         const char* addr = kLocalInetAddress; | 
 |                         auto aiStart = InetSocketAddress::getAddrInfo(addr, port); | 
 |                         if (aiStart == nullptr) return base::unique_fd{}; | 
 |                         for (auto ai = aiStart.get(); ai != nullptr; ai = ai->ai_next) { | 
 |                             auto fd = connectTo( | 
 |                                     InetSocketAddress(ai->ai_addr, ai->ai_addrlen, addr, port)); | 
 |                             if (fd.ok()) return fd; | 
 |                         } | 
 |                         ALOGE("None of the socket address resolved for %s:%u can be connected", | 
 |                               addr, port); | 
 |                         return base::unique_fd{}; | 
 |                     }; | 
 |                 } | 
 |             } | 
 |             mFd = rpcServer->releaseServer(); | 
 |             if (!mFd.fd.ok()) return AssertionFailure() << "releaseServer returns invalid fd"; | 
 |             mCtx = newFactory(rpcSecurity, mCertVerifier, std::move(auth))->newServerCtx(); | 
 |             if (mCtx == nullptr) return AssertionFailure() << "newServerCtx"; | 
 |             mSetup = true; | 
 |             return AssertionSuccess(); | 
 |         } | 
 |         RpcTransportCtx* getCtx() const { return mCtx.get(); } | 
 |         std::shared_ptr<RpcCertificateVerifierSimple> getCertVerifier() const { | 
 |             return mCertVerifier; | 
 |         } | 
 |         ConnectToServer getConnectToServerFn() { return mConnectToServer; } | 
 |         void start() { | 
 |             LOG_ALWAYS_FATAL_IF(!mSetup, "Call Server::setup first!"); | 
 |             mThread = std::make_unique<std::thread>(&Server::run, this); | 
 |         } | 
 |  | 
 |         base::unique_fd acceptServerConnection() { | 
 |             return base::unique_fd(TEMP_FAILURE_RETRY( | 
 |                     accept4(mFd.fd.get(), nullptr, nullptr, SOCK_CLOEXEC | SOCK_NONBLOCK))); | 
 |         } | 
 |  | 
 |         base::unique_fd recvmsgServerConnection() { | 
 |             std::vector<std::variant<base::unique_fd, base::borrowed_fd>> fds; | 
 |             int buf; | 
 |             iovec iov{&buf, sizeof(buf)}; | 
 |  | 
 |             if (receiveMessageFromSocket(mFd, &iov, 1, &fds) < 0) { | 
 |                 int savedErrno = errno; | 
 |                 LOG(FATAL) << "Failed receiveMessage: " << strerror(savedErrno); | 
 |             } | 
 |             if (fds.size() != 1) { | 
 |                 LOG(FATAL) << "Expected one FD from receiveMessage(), got " << fds.size(); | 
 |             } | 
 |             return std::move(std::get<base::unique_fd>(fds[0])); | 
 |         } | 
 |  | 
 |         void run() { | 
 |             LOG_ALWAYS_FATAL_IF(!mSetup, "Call Server::setup first!"); | 
 |  | 
 |             std::vector<std::thread> threads; | 
 |             while (OK == mFdTrigger->triggerablePoll(mFd, POLLIN)) { | 
 |                 base::unique_fd acceptedFd = mAcceptConnection(this); | 
 |                 threads.emplace_back(&Server::handleOne, this, std::move(acceptedFd)); | 
 |             } | 
 |  | 
 |             for (auto& thread : threads) thread.join(); | 
 |         } | 
 |         void handleOne(android::base::unique_fd acceptedFd) { | 
 |             ASSERT_TRUE(acceptedFd.ok()); | 
 |             RpcTransportFd transportFd(std::move(acceptedFd)); | 
 |             auto serverTransport = mCtx->newTransport(std::move(transportFd), mFdTrigger.get()); | 
 |             if (serverTransport == nullptr) return; // handshake failed | 
 |             ASSERT_TRUE(mPostConnect(serverTransport.get(), mFdTrigger.get())); | 
 |         } | 
 |         void shutdownAndWait() { | 
 |             shutdown(); | 
 |             join(); | 
 |         } | 
 |         void shutdown() { mFdTrigger->trigger(); } | 
 |  | 
 |         void setPostConnect( | 
 |                 std::function<AssertionResult(RpcTransport*, FdTrigger* fdTrigger)> fn) { | 
 |             mPostConnect = std::move(fn); | 
 |         } | 
 |  | 
 |     private: | 
 |         std::unique_ptr<std::thread> mThread; | 
 |         ConnectToServer mConnectToServer; | 
 |         AcceptConnection mAcceptConnection = &Server::acceptServerConnection; | 
 |         std::unique_ptr<FdTrigger> mFdTrigger = FdTrigger::make(); | 
 |         RpcTransportFd mFd, mBootstrapSocket; | 
 |         std::unique_ptr<RpcTransportCtx> mCtx; | 
 |         std::shared_ptr<RpcCertificateVerifierSimple> mCertVerifier = | 
 |                 std::make_shared<RpcCertificateVerifierSimple>(); | 
 |         bool mSetup = false; | 
 |         // The function invoked after connection and handshake. By default, it is | 
 |         // |defaultPostConnect| that sends |kMessage| to the client. | 
 |         std::function<AssertionResult(RpcTransport*, FdTrigger* fdTrigger)> mPostConnect = | 
 |                 Server::defaultPostConnect; | 
 |  | 
 |         void join() { | 
 |             if (mThread != nullptr) { | 
 |                 mThread->join(); | 
 |                 mThread = nullptr; | 
 |             } | 
 |         } | 
 |  | 
 |         static AssertionResult defaultPostConnect(RpcTransport* serverTransport, | 
 |                                                   FdTrigger* fdTrigger) { | 
 |             std::string message(kMessage); | 
 |             iovec messageIov{message.data(), message.size()}; | 
 |             auto status = serverTransport->interruptableWriteFully(fdTrigger, &messageIov, 1, | 
 |                                                                    std::nullopt, nullptr); | 
 |             if (status != OK) return AssertionFailure() << statusToString(status); | 
 |             return AssertionSuccess(); | 
 |         } | 
 |     }; | 
 |  | 
 |     class Client { | 
 |     public: | 
 |         explicit Client(ConnectToServer connectToServer) : mConnectToServer(connectToServer) {} | 
 |         Client(Client&&) = default; | 
 |         [[nodiscard]] AssertionResult setUp(const Param& param) { | 
 |             auto [socketType, rpcSecurity, certificateFormat, serverVersion] = param; | 
 |             (void)serverVersion; | 
 |             mFdTrigger = FdTrigger::make(); | 
 |             mCtx = newFactory(rpcSecurity, mCertVerifier)->newClientCtx(); | 
 |             if (mCtx == nullptr) return AssertionFailure() << "newClientCtx"; | 
 |             return AssertionSuccess(); | 
 |         } | 
 |         RpcTransportCtx* getCtx() const { return mCtx.get(); } | 
 |         std::shared_ptr<RpcCertificateVerifierSimple> getCertVerifier() const { | 
 |             return mCertVerifier; | 
 |         } | 
 |         // connect() and do handshake | 
 |         bool setUpTransport() { | 
 |             mFd = mConnectToServer(); | 
 |             if (!mFd.fd.ok()) return AssertionFailure() << "Cannot connect to server"; | 
 |             mClientTransport = mCtx->newTransport(std::move(mFd), mFdTrigger.get()); | 
 |             return mClientTransport != nullptr; | 
 |         } | 
 |         AssertionResult readMessage(const std::string& expectedMessage = kMessage) { | 
 |             LOG_ALWAYS_FATAL_IF(mClientTransport == nullptr, "setUpTransport not called or failed"); | 
 |             std::string readMessage(expectedMessage.size(), '\0'); | 
 |             iovec readMessageIov{readMessage.data(), readMessage.size()}; | 
 |             status_t readStatus = | 
 |                     mClientTransport->interruptableReadFully(mFdTrigger.get(), &readMessageIov, 1, | 
 |                                                              std::nullopt, nullptr); | 
 |             if (readStatus != OK) { | 
 |                 return AssertionFailure() << statusToString(readStatus); | 
 |             } | 
 |             if (readMessage != expectedMessage) { | 
 |                 return AssertionFailure() | 
 |                         << "Expected " << expectedMessage << ", actual " << readMessage; | 
 |             } | 
 |             return AssertionSuccess(); | 
 |         } | 
 |         void run(bool handshakeOk = true, bool readOk = true) { | 
 |             if (!setUpTransport()) { | 
 |                 ASSERT_FALSE(handshakeOk) << "newTransport returns nullptr, but it shouldn't"; | 
 |                 return; | 
 |             } | 
 |             ASSERT_TRUE(handshakeOk) << "newTransport does not return nullptr, but it should"; | 
 |             ASSERT_EQ(readOk, readMessage()); | 
 |         } | 
 |  | 
 |         bool isTransportWaiting() { return mClientTransport->isWaiting(); } | 
 |  | 
 |     private: | 
 |         ConnectToServer mConnectToServer; | 
 |         RpcTransportFd mFd; | 
 |         std::unique_ptr<FdTrigger> mFdTrigger = FdTrigger::make(); | 
 |         std::unique_ptr<RpcTransportCtx> mCtx; | 
 |         std::shared_ptr<RpcCertificateVerifierSimple> mCertVerifier = | 
 |                 std::make_shared<RpcCertificateVerifierSimple>(); | 
 |         std::unique_ptr<RpcTransport> mClientTransport; | 
 |     }; | 
 |  | 
 |     // Make A trust B. | 
 |     template <typename A, typename B> | 
 |     static status_t trust(RpcSecurity rpcSecurity, | 
 |                           std::optional<RpcCertificateFormat> certificateFormat, const A& a, | 
 |                           const B& b) { | 
 |         if (rpcSecurity != RpcSecurity::TLS) return OK; | 
 |         LOG_ALWAYS_FATAL_IF(!certificateFormat.has_value()); | 
 |         auto bCert = b->getCtx()->getCertificate(*certificateFormat); | 
 |         return a->getCertVerifier()->addTrustedPeerCertificate(*certificateFormat, bCert); | 
 |     } | 
 |  | 
 |     static constexpr const char* kMessage = "hello"; | 
 | }; | 
 |  | 
 | class RpcTransportTest : public testing::TestWithParam<RpcTransportTestUtils::Param> { | 
 | public: | 
 |     using Server = RpcTransportTestUtils::Server; | 
 |     using Client = RpcTransportTestUtils::Client; | 
 |     static inline std::string PrintParamInfo(const testing::TestParamInfo<ParamType>& info) { | 
 |         auto [socketType, rpcSecurity, certificateFormat, serverVersion] = info.param; | 
 |         auto ret = PrintToString(socketType) + "_" + newFactory(rpcSecurity)->toCString(); | 
 |         if (certificateFormat.has_value()) ret += "_" + PrintToString(*certificateFormat); | 
 |         ret += "_serverV" + std::to_string(serverVersion); | 
 |         return ret; | 
 |     } | 
 |     static std::vector<ParamType> getRpcTranportTestParams() { | 
 |         std::vector<ParamType> ret; | 
 |         for (auto serverVersion : testVersions()) { | 
 |             for (auto socketType : testSocketTypes(false /* hasPreconnected */)) { | 
 |                 for (auto rpcSecurity : RpcSecurityValues()) { | 
 |                     switch (rpcSecurity) { | 
 |                         case RpcSecurity::RAW: { | 
 |                             ret.emplace_back(socketType, rpcSecurity, std::nullopt, serverVersion); | 
 |                         } break; | 
 |                         case RpcSecurity::TLS: { | 
 |                             ret.emplace_back(socketType, rpcSecurity, RpcCertificateFormat::PEM, | 
 |                                              serverVersion); | 
 |                             ret.emplace_back(socketType, rpcSecurity, RpcCertificateFormat::DER, | 
 |                                              serverVersion); | 
 |                         } break; | 
 |                     } | 
 |                 } | 
 |             } | 
 |         } | 
 |         return ret; | 
 |     } | 
 |     template <typename A, typename B> | 
 |     status_t trust(const A& a, const B& b) { | 
 |         auto [socketType, rpcSecurity, certificateFormat, serverVersion] = GetParam(); | 
 |         (void)serverVersion; | 
 |         return RpcTransportTestUtils::trust(rpcSecurity, certificateFormat, a, b); | 
 |     } | 
 |     void SetUp() override { | 
 |         if constexpr (!kEnableRpcThreads) { | 
 |             GTEST_SKIP() << "Test skipped because threads were disabled at build time"; | 
 |         } | 
 |     } | 
 | }; | 
 |  | 
 | TEST_P(RpcTransportTest, GoodCertificate) { | 
 |     auto server = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(server->setUp(GetParam())); | 
 |  | 
 |     Client client(server->getConnectToServerFn()); | 
 |     ASSERT_TRUE(client.setUp(GetParam())); | 
 |  | 
 |     ASSERT_EQ(OK, trust(&client, server)); | 
 |     ASSERT_EQ(OK, trust(server, &client)); | 
 |  | 
 |     server->start(); | 
 |     client.run(); | 
 | } | 
 |  | 
 | TEST_P(RpcTransportTest, MultipleClients) { | 
 |     auto server = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(server->setUp(GetParam())); | 
 |  | 
 |     std::vector<Client> clients; | 
 |     for (int i = 0; i < 2; i++) { | 
 |         auto& client = clients.emplace_back(server->getConnectToServerFn()); | 
 |         ASSERT_TRUE(client.setUp(GetParam())); | 
 |         ASSERT_EQ(OK, trust(&client, server)); | 
 |         ASSERT_EQ(OK, trust(server, &client)); | 
 |     } | 
 |  | 
 |     server->start(); | 
 |     for (auto& client : clients) client.run(); | 
 | } | 
 |  | 
 | TEST_P(RpcTransportTest, UntrustedServer) { | 
 |     auto [socketType, rpcSecurity, certificateFormat, serverVersion] = GetParam(); | 
 |     (void)serverVersion; | 
 |  | 
 |     auto untrustedServer = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(untrustedServer->setUp(GetParam())); | 
 |  | 
 |     Client client(untrustedServer->getConnectToServerFn()); | 
 |     ASSERT_TRUE(client.setUp(GetParam())); | 
 |  | 
 |     ASSERT_EQ(OK, trust(untrustedServer, &client)); | 
 |  | 
 |     untrustedServer->start(); | 
 |  | 
 |     // For TLS, this should reject the certificate. For RAW sockets, it should pass because | 
 |     // the client can't verify the server's identity. | 
 |     bool handshakeOk = rpcSecurity != RpcSecurity::TLS; | 
 |     client.run(handshakeOk); | 
 | } | 
 | TEST_P(RpcTransportTest, MaliciousServer) { | 
 |     auto [socketType, rpcSecurity, certificateFormat, serverVersion] = GetParam(); | 
 |     (void)serverVersion; | 
 |  | 
 |     auto validServer = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(validServer->setUp(GetParam())); | 
 |  | 
 |     auto maliciousServer = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(maliciousServer->setUp(GetParam())); | 
 |  | 
 |     Client client(maliciousServer->getConnectToServerFn()); | 
 |     ASSERT_TRUE(client.setUp(GetParam())); | 
 |  | 
 |     ASSERT_EQ(OK, trust(&client, validServer)); | 
 |     ASSERT_EQ(OK, trust(validServer, &client)); | 
 |     ASSERT_EQ(OK, trust(maliciousServer, &client)); | 
 |  | 
 |     maliciousServer->start(); | 
 |  | 
 |     // For TLS, this should reject the certificate. For RAW sockets, it should pass because | 
 |     // the client can't verify the server's identity. | 
 |     bool handshakeOk = rpcSecurity != RpcSecurity::TLS; | 
 |     client.run(handshakeOk); | 
 | } | 
 |  | 
 | TEST_P(RpcTransportTest, UntrustedClient) { | 
 |     auto [socketType, rpcSecurity, certificateFormat, serverVersion] = GetParam(); | 
 |     (void)serverVersion; | 
 |  | 
 |     auto server = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(server->setUp(GetParam())); | 
 |  | 
 |     Client client(server->getConnectToServerFn()); | 
 |     ASSERT_TRUE(client.setUp(GetParam())); | 
 |  | 
 |     ASSERT_EQ(OK, trust(&client, server)); | 
 |  | 
 |     server->start(); | 
 |  | 
 |     // For TLS, Client should be able to verify server's identity, so client should see | 
 |     // do_handshake() successfully executed. However, server shouldn't be able to verify client's | 
 |     // identity and should drop the connection, so client shouldn't be able to read anything. | 
 |     bool readOk = rpcSecurity != RpcSecurity::TLS; | 
 |     client.run(true, readOk); | 
 | } | 
 |  | 
 | TEST_P(RpcTransportTest, MaliciousClient) { | 
 |     auto [socketType, rpcSecurity, certificateFormat, serverVersion] = GetParam(); | 
 |     (void)serverVersion; | 
 |  | 
 |     auto server = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(server->setUp(GetParam())); | 
 |  | 
 |     Client validClient(server->getConnectToServerFn()); | 
 |     ASSERT_TRUE(validClient.setUp(GetParam())); | 
 |     Client maliciousClient(server->getConnectToServerFn()); | 
 |     ASSERT_TRUE(maliciousClient.setUp(GetParam())); | 
 |  | 
 |     ASSERT_EQ(OK, trust(&validClient, server)); | 
 |     ASSERT_EQ(OK, trust(&maliciousClient, server)); | 
 |  | 
 |     server->start(); | 
 |  | 
 |     // See UntrustedClient. | 
 |     bool readOk = rpcSecurity != RpcSecurity::TLS; | 
 |     maliciousClient.run(true, readOk); | 
 | } | 
 |  | 
 | TEST_P(RpcTransportTest, Trigger) { | 
 |     std::string msg2 = ", world!"; | 
 |     std::mutex writeMutex; | 
 |     std::condition_variable writeCv; | 
 |     bool shouldContinueWriting = false; | 
 |     auto serverPostConnect = [&](RpcTransport* serverTransport, FdTrigger* fdTrigger) { | 
 |         std::string message(RpcTransportTestUtils::kMessage); | 
 |         iovec messageIov{message.data(), message.size()}; | 
 |         auto status = serverTransport->interruptableWriteFully(fdTrigger, &messageIov, 1, | 
 |                                                                std::nullopt, nullptr); | 
 |         if (status != OK) return AssertionFailure() << statusToString(status); | 
 |  | 
 |         { | 
 |             std::unique_lock<std::mutex> lock(writeMutex); | 
 |             if (!writeCv.wait_for(lock, 3s, [&] { return shouldContinueWriting; })) { | 
 |                 return AssertionFailure() << "write barrier not cleared in time!"; | 
 |             } | 
 |         } | 
 |  | 
 |         iovec msg2Iov{msg2.data(), msg2.size()}; | 
 |         status = serverTransport->interruptableWriteFully(fdTrigger, &msg2Iov, 1, std::nullopt, | 
 |                                                           nullptr); | 
 |         if (status != DEAD_OBJECT) | 
 |             return AssertionFailure() << "When FdTrigger is shut down, interruptableWriteFully " | 
 |                                          "should return DEAD_OBJECT, but it is " | 
 |                                       << statusToString(status); | 
 |         return AssertionSuccess(); | 
 |     }; | 
 |  | 
 |     auto server = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(server->setUp(GetParam())); | 
 |  | 
 |     // Set up client | 
 |     Client client(server->getConnectToServerFn()); | 
 |     ASSERT_TRUE(client.setUp(GetParam())); | 
 |  | 
 |     // Exchange keys | 
 |     ASSERT_EQ(OK, trust(&client, server)); | 
 |     ASSERT_EQ(OK, trust(server, &client)); | 
 |  | 
 |     server->setPostConnect(serverPostConnect); | 
 |  | 
 |     server->start(); | 
 |     // connect() to server and do handshake | 
 |     ASSERT_TRUE(client.setUpTransport()); | 
 |     // read the first message. This ensures that server has finished handshake and start handling | 
 |     // client fd. Server thread should pause at writeCv.wait_for(). | 
 |     ASSERT_TRUE(client.readMessage(RpcTransportTestUtils::kMessage)); | 
 |     // Trigger server shutdown after server starts handling client FD. This ensures that the second | 
 |     // write is on an FdTrigger that has been shut down. | 
 |     server->shutdown(); | 
 |     // Continues server thread to write the second message. | 
 |     { | 
 |         std::lock_guard<std::mutex> lock(writeMutex); | 
 |         shouldContinueWriting = true; | 
 |     } | 
 |     writeCv.notify_all(); | 
 |     // After this line, server thread unblocks and attempts to write the second message, but | 
 |     // shutdown is triggered, so write should failed with DEAD_OBJECT. See |serverPostConnect|. | 
 |     // On the client side, second read fails with DEAD_OBJECT | 
 |     ASSERT_FALSE(client.readMessage(msg2)); | 
 | } | 
 |  | 
 | TEST_P(RpcTransportTest, CheckWaitingForRead) { | 
 |     std::mutex readMutex; | 
 |     std::condition_variable readCv; | 
 |     bool shouldContinueReading = false; | 
 |     // Server will write data on transport once its started | 
 |     auto serverPostConnect = [&](RpcTransport* serverTransport, FdTrigger* fdTrigger) { | 
 |         std::string message(RpcTransportTestUtils::kMessage); | 
 |         iovec messageIov{message.data(), message.size()}; | 
 |         auto status = serverTransport->interruptableWriteFully(fdTrigger, &messageIov, 1, | 
 |                                                                std::nullopt, nullptr); | 
 |         if (status != OK) return AssertionFailure() << statusToString(status); | 
 |  | 
 |         { | 
 |             std::unique_lock<std::mutex> lock(readMutex); | 
 |             shouldContinueReading = true; | 
 |             lock.unlock(); | 
 |             readCv.notify_all(); | 
 |         } | 
 |         return AssertionSuccess(); | 
 |     }; | 
 |  | 
 |     // Setup Server and client | 
 |     auto server = std::make_unique<Server>(); | 
 |     ASSERT_TRUE(server->setUp(GetParam())); | 
 |  | 
 |     Client client(server->getConnectToServerFn()); | 
 |     ASSERT_TRUE(client.setUp(GetParam())); | 
 |  | 
 |     ASSERT_EQ(OK, trust(&client, server)); | 
 |     ASSERT_EQ(OK, trust(server, &client)); | 
 |     server->setPostConnect(serverPostConnect); | 
 |  | 
 |     server->start(); | 
 |     ASSERT_TRUE(client.setUpTransport()); | 
 |     { | 
 |         // Wait till server writes data | 
 |         std::unique_lock<std::mutex> lock(readMutex); | 
 |         ASSERT_TRUE(readCv.wait_for(lock, 3s, [&] { return shouldContinueReading; })); | 
 |     } | 
 |  | 
 |     // Since there is no read polling here, we will get polling count 0 | 
 |     ASSERT_FALSE(client.isTransportWaiting()); | 
 |     ASSERT_TRUE(client.readMessage(RpcTransportTestUtils::kMessage)); | 
 |     // Thread should increment polling count, read and decrement polling count | 
 |     // Again, polling count should be zero here | 
 |     ASSERT_FALSE(client.isTransportWaiting()); | 
 |  | 
 |     server->shutdown(); | 
 | } | 
 |  | 
 | INSTANTIATE_TEST_CASE_P(BinderRpc, RpcTransportTest, | 
 |                         ::testing::ValuesIn(RpcTransportTest::getRpcTranportTestParams()), | 
 |                         RpcTransportTest::PrintParamInfo); | 
 |  | 
 | class RpcTransportTlsKeyTest | 
 |       : public testing::TestWithParam< | 
 |                 std::tuple<SocketType, RpcCertificateFormat, RpcKeyFormat, uint32_t>> { | 
 | public: | 
 |     template <typename A, typename B> | 
 |     status_t trust(const A& a, const B& b) { | 
 |         auto [socketType, certificateFormat, keyFormat, serverVersion] = GetParam(); | 
 |         (void)serverVersion; | 
 |         return RpcTransportTestUtils::trust(RpcSecurity::TLS, certificateFormat, a, b); | 
 |     } | 
 |     static std::string PrintParamInfo(const testing::TestParamInfo<ParamType>& info) { | 
 |         auto [socketType, certificateFormat, keyFormat, serverVersion] = info.param; | 
 |         return PrintToString(socketType) + "_certificate_" + PrintToString(certificateFormat) + | 
 |                 "_key_" + PrintToString(keyFormat) + "_serverV" + std::to_string(serverVersion); | 
 |     }; | 
 | }; | 
 |  | 
 | TEST_P(RpcTransportTlsKeyTest, PreSignedCertificate) { | 
 |     if constexpr (!kEnableRpcThreads) { | 
 |         GTEST_SKIP() << "Test skipped because threads were disabled at build time"; | 
 |     } | 
 |  | 
 |     auto [socketType, certificateFormat, keyFormat, serverVersion] = GetParam(); | 
 |  | 
 |     std::vector<uint8_t> pkeyData, certData; | 
 |     { | 
 |         auto pkey = makeKeyPairForSelfSignedCert(); | 
 |         ASSERT_NE(nullptr, pkey); | 
 |         auto cert = makeSelfSignedCert(pkey.get(), kCertValidSeconds); | 
 |         ASSERT_NE(nullptr, cert); | 
 |         pkeyData = serializeUnencryptedPrivatekey(pkey.get(), keyFormat); | 
 |         certData = serializeCertificate(cert.get(), certificateFormat); | 
 |     } | 
 |  | 
 |     auto desPkey = deserializeUnencryptedPrivatekey(pkeyData, keyFormat); | 
 |     auto desCert = deserializeCertificate(certData, certificateFormat); | 
 |     auto auth = std::make_unique<RpcAuthPreSigned>(std::move(desPkey), std::move(desCert)); | 
 |     auto utilsParam = std::make_tuple(socketType, RpcSecurity::TLS, | 
 |                                       std::make_optional(certificateFormat), serverVersion); | 
 |  | 
 |     auto server = std::make_unique<RpcTransportTestUtils::Server>(); | 
 |     ASSERT_TRUE(server->setUp(utilsParam, std::move(auth))); | 
 |  | 
 |     RpcTransportTestUtils::Client client(server->getConnectToServerFn()); | 
 |     ASSERT_TRUE(client.setUp(utilsParam)); | 
 |  | 
 |     ASSERT_EQ(OK, trust(&client, server)); | 
 |     ASSERT_EQ(OK, trust(server, &client)); | 
 |  | 
 |     server->start(); | 
 |     client.run(); | 
 | } | 
 |  | 
 | INSTANTIATE_TEST_CASE_P( | 
 |         BinderRpc, RpcTransportTlsKeyTest, | 
 |         testing::Combine(testing::ValuesIn(testSocketTypes(false /* hasPreconnected*/)), | 
 |                          testing::Values(RpcCertificateFormat::PEM, RpcCertificateFormat::DER), | 
 |                          testing::Values(RpcKeyFormat::PEM, RpcKeyFormat::DER), | 
 |                          testing::ValuesIn(testVersions())), | 
 |         RpcTransportTlsKeyTest::PrintParamInfo); | 
 |  | 
 | } // namespace android | 
 |  | 
 | int main(int argc, char** argv) { | 
 |     ::testing::InitGoogleTest(&argc, argv); | 
 |     android::base::InitLogging(argv, android::base::StderrLogger, android::base::DefaultAborter); | 
 |  | 
 |     return RUN_ALL_TESTS(); | 
 | } |