blob: 0bae95de87cb717971a8ba1360ef0207bdae4c88 [file] [log] [blame]
Michael Butler4b276a72020-08-06 23:22:35 -07001/*
2 * Copyright (C) 2020 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17#include "PreparedModel.h"
18
19#include "Callbacks.h"
20#include "Conversions.h"
21#include "Utils.h"
22
23#include <android/hardware/neuralnetworks/1.0/types.h>
24#include <android/hardware/neuralnetworks/1.1/types.h>
25#include <android/hardware/neuralnetworks/1.2/types.h>
26#include <android/hardware/neuralnetworks/1.3/IPreparedModel.h>
27#include <android/hardware/neuralnetworks/1.3/types.h>
28#include <nnapi/IPreparedModel.h>
29#include <nnapi/Result.h>
Michael Butler6547b2a2020-11-22 19:36:30 -080030#include <nnapi/TypeUtils.h>
Michael Butler4b276a72020-08-06 23:22:35 -070031#include <nnapi/Types.h>
32#include <nnapi/hal/1.2/Conversions.h>
33#include <nnapi/hal/CommonUtils.h>
34#include <nnapi/hal/HandleError.h>
35#include <nnapi/hal/ProtectCallback.h>
36
37#include <memory>
38#include <tuple>
39#include <utility>
40#include <vector>
41
Michael Butleraad934b2020-12-13 23:06:06 -080042// See hardware/interfaces/neuralnetworks/utils/README.md for more information on HIDL interface
43// lifetimes across processes and for protecting asynchronous calls across HIDL.
44
Michael Butler4b276a72020-08-06 23:22:35 -070045namespace android::hardware::neuralnetworks::V1_3::utils {
46namespace {
47
48nn::GeneralResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>>
49convertExecutionResultsHelper(const hidl_vec<V1_2::OutputShape>& outputShapes,
50 const V1_2::Timing& timing) {
Michael Butler6547b2a2020-11-22 19:36:30 -080051 return std::make_pair(NN_TRY(nn::convert(outputShapes)), NN_TRY(nn::convert(timing)));
Michael Butler4b276a72020-08-06 23:22:35 -070052}
53
54nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>> convertExecutionResults(
55 const hidl_vec<V1_2::OutputShape>& outputShapes, const V1_2::Timing& timing) {
56 return hal::utils::makeExecutionFailure(convertExecutionResultsHelper(outputShapes, timing));
57}
58
Michael Butler4b276a72020-08-06 23:22:35 -070059nn::GeneralResult<std::pair<nn::Timing, nn::Timing>> convertFencedExecutionCallbackResults(
60 const V1_2::Timing& timingLaunched, const V1_2::Timing& timingFenced) {
Michael Butler6547b2a2020-11-22 19:36:30 -080061 return std::make_pair(NN_TRY(nn::convert(timingLaunched)), NN_TRY(nn::convert(timingFenced)));
Michael Butler4b276a72020-08-06 23:22:35 -070062}
63
64nn::GeneralResult<std::pair<nn::SyncFence, nn::ExecuteFencedInfoCallback>>
65convertExecuteFencedResults(const hidl_handle& syncFence,
66 const sp<IFencedExecutionCallback>& callback) {
67 auto resultSyncFence = nn::SyncFence::createAsSignaled();
68 if (syncFence.getNativeHandle() != nullptr) {
Michael Butler6547b2a2020-11-22 19:36:30 -080069 auto sharedHandle = NN_TRY(nn::convert(syncFence));
Michael Butler4b276a72020-08-06 23:22:35 -070070 resultSyncFence = NN_TRY(hal::utils::makeGeneralFailure(
Michael Butler6547b2a2020-11-22 19:36:30 -080071 nn::SyncFence::create(std::move(sharedHandle)), nn::ErrorStatus::GENERAL_FAILURE));
Michael Butler4b276a72020-08-06 23:22:35 -070072 }
73
74 if (callback == nullptr) {
75 return NN_ERROR(nn::ErrorStatus::GENERAL_FAILURE) << "callback is null";
76 }
77
78 // Create callback which can be used to retrieve the execution error status and timings.
79 nn::ExecuteFencedInfoCallback resultCallback =
80 [callback]() -> nn::GeneralResult<std::pair<nn::Timing, nn::Timing>> {
81 nn::GeneralResult<std::pair<nn::Timing, nn::Timing>> result =
82 NN_ERROR(nn::ErrorStatus::GENERAL_FAILURE) << "uninitialized";
83 auto cb = [&result](ErrorStatus status, const V1_2::Timing& timingLaunched,
84 const V1_2::Timing& timingFenced) {
85 if (status != ErrorStatus::NONE) {
Michael Butler6547b2a2020-11-22 19:36:30 -080086 const auto canonical =
87 nn::convert(status).value_or(nn::ErrorStatus::GENERAL_FAILURE);
Michael Butler4b276a72020-08-06 23:22:35 -070088 result = NN_ERROR(canonical) << "getExecutionInfo failed with " << toString(status);
89 } else {
90 result = convertFencedExecutionCallbackResults(timingLaunched, timingFenced);
91 }
92 };
93
94 const auto ret = callback->getExecutionInfo(cb);
Michael Butlercca3e202020-11-22 20:25:34 -080095 HANDLE_TRANSPORT_FAILURE(ret);
Michael Butler4b276a72020-08-06 23:22:35 -070096
97 return result;
98 };
99
100 return std::make_pair(std::move(resultSyncFence), std::move(resultCallback));
101}
102
103} // namespace
104
105nn::GeneralResult<std::shared_ptr<const PreparedModel>> PreparedModel::create(
106 sp<V1_3::IPreparedModel> preparedModel) {
107 if (preparedModel == nullptr) {
108 return NN_ERROR(nn::ErrorStatus::INVALID_ARGUMENT)
109 << "V1_3::utils::PreparedModel::create must have non-null preparedModel";
110 }
111
112 auto deathHandler = NN_TRY(hal::utils::DeathHandler::create(preparedModel));
113 return std::make_shared<const PreparedModel>(PrivateConstructorTag{}, std::move(preparedModel),
114 std::move(deathHandler));
115}
116
117PreparedModel::PreparedModel(PrivateConstructorTag /*tag*/, sp<V1_3::IPreparedModel> preparedModel,
118 hal::utils::DeathHandler deathHandler)
119 : kPreparedModel(std::move(preparedModel)), kDeathHandler(std::move(deathHandler)) {}
120
121nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>>
122PreparedModel::executeSynchronously(const Request& request, V1_2::MeasureTiming measure,
123 const OptionalTimePoint& deadline,
124 const OptionalTimeoutDuration& loopTimeoutDuration) const {
125 nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>> result =
126 NN_ERROR(nn::ErrorStatus::GENERAL_FAILURE) << "uninitialized";
127 const auto cb = [&result](ErrorStatus status, const hidl_vec<V1_2::OutputShape>& outputShapes,
128 const V1_2::Timing& timing) {
129 if (status != ErrorStatus::NONE) {
Michael Butler6547b2a2020-11-22 19:36:30 -0800130 const auto canonical = nn::convert(status).value_or(nn::ErrorStatus::GENERAL_FAILURE);
Michael Butler4b276a72020-08-06 23:22:35 -0700131 result = NN_ERROR(canonical) << "executeSynchronously failed with " << toString(status);
132 } else {
133 result = convertExecutionResults(outputShapes, timing);
134 }
135 };
136
137 const auto ret = kPreparedModel->executeSynchronously_1_3(request, measure, deadline,
138 loopTimeoutDuration, cb);
Michael Butlercca3e202020-11-22 20:25:34 -0800139 HANDLE_TRANSPORT_FAILURE(ret);
Michael Butler4b276a72020-08-06 23:22:35 -0700140
141 return result;
142}
143
144nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>>
145PreparedModel::executeAsynchronously(const Request& request, V1_2::MeasureTiming measure,
146 const OptionalTimePoint& deadline,
147 const OptionalTimeoutDuration& loopTimeoutDuration) const {
148 const auto cb = sp<ExecutionCallback>::make();
149 const auto scoped = kDeathHandler.protectCallback(cb.get());
150
151 const auto ret =
152 kPreparedModel->execute_1_3(request, measure, deadline, loopTimeoutDuration, cb);
Michael Butlercca3e202020-11-22 20:25:34 -0800153 const auto status = HANDLE_TRANSPORT_FAILURE(ret);
Michael Butler4b276a72020-08-06 23:22:35 -0700154 if (status != ErrorStatus::NONE) {
Michael Butler6547b2a2020-11-22 19:36:30 -0800155 const auto canonical = nn::convert(status).value_or(nn::ErrorStatus::GENERAL_FAILURE);
Michael Butler4b276a72020-08-06 23:22:35 -0700156 return NN_ERROR(canonical) << "executeAsynchronously failed with " << toString(status);
157 }
158
159 return cb->get();
160}
161
162nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>> PreparedModel::execute(
163 const nn::Request& request, nn::MeasureTiming measure,
164 const nn::OptionalTimePoint& deadline,
Michael Butler4024d8f2020-12-04 17:38:20 -0800165 const nn::OptionalDuration& loopTimeoutDuration) const {
Michael Butler4b276a72020-08-06 23:22:35 -0700166 // Ensure that request is ready for IPC.
167 std::optional<nn::Request> maybeRequestInShared;
168 const nn::Request& requestInShared = NN_TRY(hal::utils::makeExecutionFailure(
169 hal::utils::flushDataFromPointerToShared(&request, &maybeRequestInShared)));
170
171 const auto hidlRequest = NN_TRY(hal::utils::makeExecutionFailure(convert(requestInShared)));
172 const auto hidlMeasure =
173 NN_TRY(hal::utils::makeExecutionFailure(V1_2::utils::convert(measure)));
174 const auto hidlDeadline = NN_TRY(hal::utils::makeExecutionFailure(convert(deadline)));
175 const auto hidlLoopTimeoutDuration =
176 NN_TRY(hal::utils::makeExecutionFailure(convert(loopTimeoutDuration)));
177
178 nn::ExecutionResult<std::pair<std::vector<nn::OutputShape>, nn::Timing>> result =
179 NN_ERROR(nn::ErrorStatus::GENERAL_FAILURE) << "uninitialized";
180 const bool preferSynchronous = true;
181
182 // Execute synchronously if allowed.
183 if (preferSynchronous) {
184 result = executeSynchronously(hidlRequest, hidlMeasure, hidlDeadline,
185 hidlLoopTimeoutDuration);
186 }
187
188 // Run asymchronous execution if execution has not already completed.
189 if (!result.has_value()) {
190 result = executeAsynchronously(hidlRequest, hidlMeasure, hidlDeadline,
191 hidlLoopTimeoutDuration);
192 }
193
194 // Flush output buffers if suxcessful execution.
195 if (result.has_value()) {
196 NN_TRY(hal::utils::makeExecutionFailure(
197 hal::utils::unflushDataFromSharedToPointer(request, maybeRequestInShared)));
198 }
199
200 return result;
201}
202
203nn::GeneralResult<std::pair<nn::SyncFence, nn::ExecuteFencedInfoCallback>>
204PreparedModel::executeFenced(const nn::Request& request, const std::vector<nn::SyncFence>& waitFor,
205 nn::MeasureTiming measure, const nn::OptionalTimePoint& deadline,
Michael Butler4024d8f2020-12-04 17:38:20 -0800206 const nn::OptionalDuration& loopTimeoutDuration,
207 const nn::OptionalDuration& timeoutDurationAfterFence) const {
Michael Butler4b276a72020-08-06 23:22:35 -0700208 // Ensure that request is ready for IPC.
209 std::optional<nn::Request> maybeRequestInShared;
210 const nn::Request& requestInShared =
211 NN_TRY(hal::utils::flushDataFromPointerToShared(&request, &maybeRequestInShared));
212
213 const auto hidlRequest = NN_TRY(convert(requestInShared));
Slava Shklyaev49817a02020-10-27 18:44:01 +0000214 const auto hidlWaitFor = NN_TRY(hal::utils::convertSyncFences(waitFor));
Michael Butler4b276a72020-08-06 23:22:35 -0700215 const auto hidlMeasure = NN_TRY(V1_2::utils::convert(measure));
216 const auto hidlDeadline = NN_TRY(convert(deadline));
217 const auto hidlLoopTimeoutDuration = NN_TRY(convert(loopTimeoutDuration));
218 const auto hidlTimeoutDurationAfterFence = NN_TRY(convert(timeoutDurationAfterFence));
219
220 nn::GeneralResult<std::pair<nn::SyncFence, nn::ExecuteFencedInfoCallback>> result =
221 NN_ERROR(nn::ErrorStatus::GENERAL_FAILURE) << "uninitialized";
222 auto cb = [&result](ErrorStatus status, const hidl_handle& syncFence,
223 const sp<IFencedExecutionCallback>& callback) {
224 if (status != ErrorStatus::NONE) {
Michael Butler6547b2a2020-11-22 19:36:30 -0800225 const auto canonical = nn::convert(status).value_or(nn::ErrorStatus::GENERAL_FAILURE);
Michael Butler4b276a72020-08-06 23:22:35 -0700226 result = NN_ERROR(canonical) << "executeFenced failed with " << toString(status);
227 } else {
228 result = convertExecuteFencedResults(syncFence, callback);
229 }
230 };
231
232 const auto ret = kPreparedModel->executeFenced(hidlRequest, hidlWaitFor, hidlMeasure,
233 hidlDeadline, hidlLoopTimeoutDuration,
234 hidlTimeoutDurationAfterFence, cb);
Michael Butlercca3e202020-11-22 20:25:34 -0800235 HANDLE_TRANSPORT_FAILURE(ret);
Michael Butler4b276a72020-08-06 23:22:35 -0700236 auto [syncFence, callback] = NN_TRY(std::move(result));
237
238 // If executeFenced required the request memory to be moved into shared memory, block here until
239 // the fenced execution has completed and flush the memory back.
240 if (maybeRequestInShared.has_value()) {
241 const auto state = syncFence.syncWait({});
242 if (state != nn::SyncFence::FenceState::SIGNALED) {
243 return NN_ERROR() << "syncWait failed with " << state;
244 }
245 NN_TRY(hal::utils::unflushDataFromSharedToPointer(request, maybeRequestInShared));
246 }
247
248 return std::make_pair(std::move(syncFence), std::move(callback));
249}
250
251std::any PreparedModel::getUnderlyingResource() const {
252 sp<V1_3::IPreparedModel> resource = kPreparedModel;
253 return resource;
254}
255
256} // namespace android::hardware::neuralnetworks::V1_3::utils