| // |
| // Copyright © 2021 Arm Ltd. All rights reserved. |
| // SPDX-License-Identifier: MIT |
| // |
| |
| #include "ClImportTensorHandleFactory.hpp" |
| #include "ClImportTensorHandle.hpp" |
| |
| #include <armnn/utility/NumericCast.hpp> |
| #include <armnn/utility/PolymorphicDowncast.hpp> |
| |
| #include <arm_compute/core/Coordinates.h> |
| #include <arm_compute/runtime/CL/CLTensor.h> |
| |
| namespace armnn |
| { |
| |
| using FactoryId = ITensorHandleFactory::FactoryId; |
| |
| std::unique_ptr<ITensorHandle> ClImportTensorHandleFactory::CreateSubTensorHandle( |
| ITensorHandle& parent, const TensorShape& subTensorShape, const unsigned int* subTensorOrigin) const |
| { |
| arm_compute::Coordinates coords; |
| arm_compute::TensorShape shape = armcomputetensorutils::BuildArmComputeTensorShape(subTensorShape); |
| |
| coords.set_num_dimensions(subTensorShape.GetNumDimensions()); |
| for (unsigned int i = 0; i < subTensorShape.GetNumDimensions(); ++i) |
| { |
| // Arm compute indexes tensor coords in reverse order. |
| unsigned int revertedIndex = subTensorShape.GetNumDimensions() - i - 1; |
| coords.set(i, armnn::numeric_cast<int>(subTensorOrigin[revertedIndex])); |
| } |
| |
| const arm_compute::TensorShape parentShape = armcomputetensorutils::BuildArmComputeTensorShape(parent.GetShape()); |
| |
| // In order for ACL to support subtensors the concat axis cannot be on x or y and the values of x and y |
| // must match the parent shapes |
| if (coords.x() != 0 || coords.y() != 0) |
| { |
| return nullptr; |
| } |
| if ((parentShape.x() != shape.x()) || (parentShape.y() != shape.y())) |
| { |
| return nullptr; |
| } |
| |
| if (!::arm_compute::error_on_invalid_subtensor(__func__, __FILE__, __LINE__, parentShape, coords, shape)) |
| { |
| return nullptr; |
| } |
| |
| return std::make_unique<ClImportSubTensorHandle>( |
| PolymorphicDowncast<IClImportTensorHandle*>(&parent), shape, coords); |
| } |
| |
| std::unique_ptr<ITensorHandle> ClImportTensorHandleFactory::CreateTensorHandle(const TensorInfo& tensorInfo) const |
| { |
| std::unique_ptr<ClImportTensorHandle> tensorHandle = std::make_unique<ClImportTensorHandle>(tensorInfo, |
| GetImportFlags()); |
| return tensorHandle; |
| } |
| |
| std::unique_ptr<ITensorHandle> ClImportTensorHandleFactory::CreateTensorHandle(const TensorInfo& tensorInfo, |
| DataLayout dataLayout) const |
| { |
| std::unique_ptr<ClImportTensorHandle> tensorHandle = std::make_unique<ClImportTensorHandle>(tensorInfo, |
| dataLayout, |
| GetImportFlags()); |
| return tensorHandle; |
| } |
| |
| std::unique_ptr<ITensorHandle> ClImportTensorHandleFactory::CreateTensorHandle(const TensorInfo& tensorInfo, |
| const bool IsMemoryManaged) const |
| { |
| if (IsMemoryManaged) |
| { |
| throw InvalidArgumentException("ClImportTensorHandleFactory does not support memory managed tensors."); |
| } |
| return CreateTensorHandle(tensorInfo); |
| } |
| |
| std::unique_ptr<ITensorHandle> ClImportTensorHandleFactory::CreateTensorHandle(const TensorInfo& tensorInfo, |
| DataLayout dataLayout, |
| const bool IsMemoryManaged) const |
| { |
| if (IsMemoryManaged) |
| { |
| throw InvalidArgumentException("ClImportTensorHandleFactory does not support memory managed tensors."); |
| } |
| return CreateTensorHandle(tensorInfo, dataLayout); |
| } |
| |
| const FactoryId& ClImportTensorHandleFactory::GetIdStatic() |
| { |
| static const FactoryId s_Id(ClImportTensorHandleFactoryId()); |
| return s_Id; |
| } |
| |
| const FactoryId& ClImportTensorHandleFactory::GetId() const |
| { |
| return GetIdStatic(); |
| } |
| |
| bool ClImportTensorHandleFactory::SupportsSubTensors() const |
| { |
| return true; |
| } |
| |
| bool ClImportTensorHandleFactory::SupportsMapUnmap() const |
| { |
| return false; |
| } |
| |
| MemorySourceFlags ClImportTensorHandleFactory::GetExportFlags() const |
| { |
| return m_ExportFlags; |
| } |
| |
| MemorySourceFlags ClImportTensorHandleFactory::GetImportFlags() const |
| { |
| return m_ImportFlags; |
| } |
| |
| std::vector<Capability> ClImportTensorHandleFactory::GetCapabilities(const IConnectableLayer* layer, |
| const IConnectableLayer* connectedLayer, |
| CapabilityClass capabilityClass) |
| { |
| IgnoreUnused(layer); |
| IgnoreUnused(connectedLayer); |
| std::vector<Capability> capabilities; |
| if (capabilityClass == CapabilityClass::FallbackImportDisabled) |
| { |
| Capability paddingCapability(CapabilityClass::FallbackImportDisabled, true); |
| capabilities.push_back(paddingCapability); |
| } |
| return capabilities; |
| } |
| |
| } // namespace armnn |