Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 1 | // |
| 2 | // Copyright © 2021 Arm Ltd and Contributors. All rights reserved. |
| 3 | // SPDX-License-Identifier: MIT |
| 4 | // |
| 5 | |
| 6 | #include <armnn/Exceptions.hpp> |
| 7 | |
James Conroy | 1f58f03 | 2021-04-27 17:13:27 +0100 | [diff] [blame] | 8 | #include <backendsCommon/TensorHandle.hpp> |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 9 | #include <backendsCommon/Workload.hpp> |
| 10 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 11 | #include <doctest/doctest.h> |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 12 | |
| 13 | using namespace armnn; |
| 14 | |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 15 | |
| 16 | namespace |
| 17 | { |
| 18 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 19 | TEST_SUITE("WorkloadAsyncExecuteTests") |
| 20 | { |
| 21 | |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 22 | struct Workload0 : BaseWorkload<ElementwiseUnaryQueueDescriptor> |
| 23 | { |
| 24 | Workload0(const ElementwiseUnaryQueueDescriptor& descriptor, const WorkloadInfo& info) |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 25 | : BaseWorkload(descriptor, info) |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 26 | { |
| 27 | } |
| 28 | |
| 29 | Workload0() : BaseWorkload(ElementwiseUnaryQueueDescriptor(), WorkloadInfo()) |
| 30 | { |
| 31 | } |
| 32 | |
| 33 | void Execute() const |
| 34 | { |
| 35 | int* inVals = static_cast<int*>(m_Data.m_Inputs[0][0].Map()); |
| 36 | int* outVals = static_cast<int*>(m_Data.m_Outputs[0][0].Map()); |
| 37 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 38 | for (unsigned int i = 0; |
| 39 | i < m_Data.m_Inputs[0][0].GetShape().GetNumElements(); |
| 40 | ++i) |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 41 | { |
| 42 | outVals[i] = inVals[i] * outVals[i]; |
| 43 | inVals[i] = outVals[i]; |
| 44 | } |
| 45 | } |
| 46 | |
| 47 | void ExecuteAsync(WorkingMemDescriptor& desc) |
| 48 | { |
| 49 | int* inVals = static_cast<int*>(desc.m_Inputs[0][0].Map()); |
| 50 | int* outVals = static_cast<int*>(desc.m_Outputs[0][0].Map()); |
| 51 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 52 | for (unsigned int i = 0; |
| 53 | i < desc.m_Inputs[0][0].GetShape().GetNumElements(); |
| 54 | ++i) |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 55 | { |
| 56 | outVals[i] = inVals[i] + outVals[i]; |
| 57 | inVals[i] = outVals[i]; |
| 58 | } |
| 59 | } |
| 60 | |
| 61 | QueueDescriptor* GetQueueDescriptor() |
| 62 | { |
| 63 | return &m_Data; |
| 64 | } |
| 65 | }; |
| 66 | |
| 67 | struct Workload1 : BaseWorkload<ElementwiseUnaryQueueDescriptor> |
| 68 | { |
| 69 | Workload1(const ElementwiseUnaryQueueDescriptor& descriptor, const WorkloadInfo& info) |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 70 | : BaseWorkload(descriptor, info) |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 71 | { |
| 72 | } |
| 73 | |
| 74 | void Execute() const |
| 75 | { |
| 76 | int* inVals = static_cast<int*>(m_Data.m_Inputs[0][0].Map()); |
| 77 | int* outVals = static_cast<int*>(m_Data.m_Outputs[0][0].Map()); |
| 78 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 79 | for (unsigned int i = 0; |
| 80 | i < m_Data.m_Inputs[0][0].GetShape().GetNumElements(); |
| 81 | ++i) |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 82 | { |
| 83 | outVals[i] = inVals[i] * outVals[i]; |
| 84 | inVals[i] = outVals[i]; |
| 85 | } |
| 86 | } |
| 87 | }; |
| 88 | |
| 89 | void ValidateTensor(ITensorHandle* tensorHandle, int expectedValue) |
| 90 | { |
| 91 | int* actualOutput = static_cast<int*>(tensorHandle->Map()); |
| 92 | |
| 93 | bool allValuesCorrect = true; |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 94 | for (unsigned int i = 0; |
| 95 | i < tensorHandle->GetShape().GetNumElements(); |
| 96 | ++i) |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 97 | { |
| 98 | if (actualOutput[i] != expectedValue) |
| 99 | { |
| 100 | allValuesCorrect = false; |
| 101 | } |
| 102 | } |
| 103 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 104 | CHECK(allValuesCorrect); |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 105 | } |
| 106 | |
| 107 | template<typename Workload> |
| 108 | std::unique_ptr<Workload> CreateWorkload(TensorInfo info, ITensorHandle* inputTensor, ITensorHandle* outputTensor) |
| 109 | { |
| 110 | WorkloadInfo workloadInfo; |
| 111 | workloadInfo.m_InputTensorInfos = std::vector<TensorInfo>{info}; |
| 112 | workloadInfo.m_OutputTensorInfos = std::vector<TensorInfo>{info}; |
| 113 | |
| 114 | ElementwiseUnaryQueueDescriptor elementwiseUnaryQueueDescriptor; |
| 115 | elementwiseUnaryQueueDescriptor.m_Inputs = std::vector<ITensorHandle*>{inputTensor}; |
| 116 | elementwiseUnaryQueueDescriptor.m_Outputs = std::vector<ITensorHandle*>{outputTensor}; |
| 117 | |
| 118 | return std::make_unique<Workload>(elementwiseUnaryQueueDescriptor, workloadInfo); |
| 119 | } |
| 120 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 121 | TEST_CASE("TestAsyncExecute") |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 122 | { |
| 123 | TensorInfo info({5}, DataType::Signed32); |
| 124 | |
| 125 | int inVals[5]{2, 2, 2, 2, 2}; |
| 126 | int outVals[5]{1, 1, 1, 1, 1}; |
| 127 | |
| 128 | int expectedExecuteval = 2; |
| 129 | int expectedExecuteAsyncval = 3; |
| 130 | |
| 131 | ConstTensor constInputTensor(info, inVals); |
| 132 | ConstTensor constOutputTensor(info, outVals); |
| 133 | |
James Conroy | 1f58f03 | 2021-04-27 17:13:27 +0100 | [diff] [blame] | 134 | ScopedTensorHandle syncInput0(constInputTensor); |
| 135 | ScopedTensorHandle syncOutput0(constOutputTensor); |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 136 | |
| 137 | std::unique_ptr<Workload0> workload0 = CreateWorkload<Workload0>(info, &syncInput0, &syncOutput0); |
| 138 | |
| 139 | workload0.get()->Execute(); |
| 140 | |
James Conroy | 1f58f03 | 2021-04-27 17:13:27 +0100 | [diff] [blame] | 141 | ScopedTensorHandle asyncInput0(constInputTensor); |
| 142 | ScopedTensorHandle asyncOutput0(constOutputTensor); |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 143 | |
| 144 | WorkingMemDescriptor workingMemDescriptor0; |
| 145 | workingMemDescriptor0.m_Inputs = std::vector<ITensorHandle*>{&asyncInput0}; |
| 146 | workingMemDescriptor0.m_Outputs = std::vector<ITensorHandle*>{&asyncOutput0}; |
| 147 | |
| 148 | workload0.get()->ExecuteAsync(workingMemDescriptor0); |
| 149 | |
| 150 | // Inputs are also changed by the execute/executeAsync calls to make sure there is no interference with them |
| 151 | ValidateTensor(workingMemDescriptor0.m_Outputs[0], expectedExecuteAsyncval); |
| 152 | ValidateTensor(workingMemDescriptor0.m_Inputs[0], expectedExecuteAsyncval); |
| 153 | |
| 154 | ValidateTensor(&workload0.get()->GetQueueDescriptor()->m_Outputs[0][0], expectedExecuteval); |
| 155 | ValidateTensor(&workload0.get()->GetQueueDescriptor()->m_Inputs[0][0], expectedExecuteval); |
| 156 | } |
| 157 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 158 | TEST_CASE("TestDefaultAsyncExecute") |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 159 | { |
| 160 | TensorInfo info({5}, DataType::Signed32); |
| 161 | |
| 162 | std::vector<int> inVals{2, 2, 2, 2, 2}; |
| 163 | std::vector<int> outVals{1, 1, 1, 1, 1}; |
| 164 | std::vector<int> defaultVals{0, 0, 0, 0, 0}; |
| 165 | |
| 166 | int expectedExecuteval = 2; |
| 167 | |
| 168 | ConstTensor constInputTensor(info, inVals); |
| 169 | ConstTensor constOutputTensor(info, outVals); |
| 170 | ConstTensor defaultTensor(info, &defaultVals); |
| 171 | |
James Conroy | 1f58f03 | 2021-04-27 17:13:27 +0100 | [diff] [blame] | 172 | ScopedTensorHandle defaultInput = ScopedTensorHandle(defaultTensor); |
| 173 | ScopedTensorHandle defaultOutput = ScopedTensorHandle(defaultTensor); |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 174 | |
| 175 | std::unique_ptr<Workload1> workload1 = CreateWorkload<Workload1>(info, &defaultInput, &defaultOutput); |
| 176 | |
James Conroy | 1f58f03 | 2021-04-27 17:13:27 +0100 | [diff] [blame] | 177 | ScopedTensorHandle asyncInput(constInputTensor); |
| 178 | ScopedTensorHandle asyncOutput(constOutputTensor); |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 179 | |
| 180 | WorkingMemDescriptor workingMemDescriptor; |
| 181 | workingMemDescriptor.m_Inputs = std::vector<ITensorHandle*>{&asyncInput}; |
| 182 | workingMemDescriptor.m_Outputs = std::vector<ITensorHandle*>{&asyncOutput}; |
| 183 | |
| 184 | workload1.get()->ExecuteAsync(workingMemDescriptor); |
| 185 | |
| 186 | // workload1 has no AsyncExecute implementation and so should use the default workload AsyncExecute |
| 187 | // implementation which will call workload1.Execute() in a thread safe manner |
| 188 | ValidateTensor(workingMemDescriptor.m_Outputs[0], expectedExecuteval); |
| 189 | ValidateTensor(workingMemDescriptor.m_Inputs[0], expectedExecuteval); |
| 190 | } |
| 191 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 192 | TEST_CASE("TestDefaultAsyncExeuteWithThreads") |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 193 | { |
| 194 | // Use a large vector so the threads have a chance to interact |
| 195 | unsigned int vecSize = 1000; |
| 196 | TensorInfo info({vecSize}, DataType::Signed32); |
| 197 | |
| 198 | std::vector<int> inVals1(vecSize, 2); |
| 199 | std::vector<int> outVals1(vecSize, 1); |
| 200 | std::vector<int> inVals2(vecSize, 5); |
| 201 | std::vector<int> outVals2(vecSize, -1); |
| 202 | |
| 203 | std::vector<int> defaultVals(vecSize, 0); |
| 204 | |
| 205 | int expectedExecuteval1 = 4; |
| 206 | int expectedExecuteval2 = 25; |
| 207 | ConstTensor constInputTensor1(info, inVals1); |
| 208 | ConstTensor constOutputTensor1(info, outVals1); |
| 209 | |
| 210 | ConstTensor constInputTensor2(info, inVals2); |
| 211 | ConstTensor constOutputTensor2(info, outVals2); |
| 212 | |
| 213 | ConstTensor defaultTensor(info, &defaultVals); |
| 214 | |
James Conroy | 1f58f03 | 2021-04-27 17:13:27 +0100 | [diff] [blame] | 215 | ScopedTensorHandle defaultInput = ScopedTensorHandle(defaultTensor); |
| 216 | ScopedTensorHandle defaultOutput = ScopedTensorHandle(defaultTensor); |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 217 | std::unique_ptr<Workload1> workload = CreateWorkload<Workload1>(info, &defaultInput, &defaultOutput); |
| 218 | |
James Conroy | 1f58f03 | 2021-04-27 17:13:27 +0100 | [diff] [blame] | 219 | ScopedTensorHandle asyncInput1(constInputTensor1); |
| 220 | ScopedTensorHandle asyncOutput1(constOutputTensor1); |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 221 | |
| 222 | WorkingMemDescriptor workingMemDescriptor1; |
| 223 | workingMemDescriptor1.m_Inputs = std::vector<ITensorHandle*>{&asyncInput1}; |
| 224 | workingMemDescriptor1.m_Outputs = std::vector<ITensorHandle*>{&asyncOutput1}; |
| 225 | |
| 226 | |
James Conroy | 1f58f03 | 2021-04-27 17:13:27 +0100 | [diff] [blame] | 227 | ScopedTensorHandle asyncInput2(constInputTensor2); |
| 228 | ScopedTensorHandle asyncOutput2(constOutputTensor2); |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 229 | |
| 230 | WorkingMemDescriptor workingMemDescriptor2; |
| 231 | workingMemDescriptor2.m_Inputs = std::vector<ITensorHandle*>{&asyncInput2}; |
| 232 | workingMemDescriptor2.m_Outputs = std::vector<ITensorHandle*>{&asyncOutput2}; |
| 233 | |
| 234 | std::thread thread1 = std::thread([&]() |
| 235 | { |
| 236 | workload.get()->ExecuteAsync(workingMemDescriptor1); |
| 237 | workload.get()->ExecuteAsync(workingMemDescriptor1); |
| 238 | }); |
| 239 | |
| 240 | std::thread thread2 = std::thread([&]() |
| 241 | { |
| 242 | workload.get()->ExecuteAsync(workingMemDescriptor2); |
| 243 | workload.get()->ExecuteAsync(workingMemDescriptor2); |
| 244 | }); |
| 245 | |
| 246 | thread1.join(); |
| 247 | thread2.join(); |
| 248 | |
| 249 | ValidateTensor(workingMemDescriptor1.m_Outputs[0], expectedExecuteval1); |
| 250 | ValidateTensor(workingMemDescriptor1.m_Inputs[0], expectedExecuteval1); |
| 251 | |
| 252 | ValidateTensor(workingMemDescriptor2.m_Outputs[0], expectedExecuteval2); |
| 253 | ValidateTensor(workingMemDescriptor2.m_Inputs[0], expectedExecuteval2); |
| 254 | } |
| 255 | |
Sadik Armagan | 1625efc | 2021-06-10 18:24:34 +0100 | [diff] [blame] | 256 | } |
Finn Williams | b76eaed | 2021-03-31 16:22:40 +0100 | [diff] [blame] | 257 | |
| 258 | } |