blob: 06fa5edde8c7ea6a5fe8d52a514086ed0e745ded [file] [log] [blame]
//
// Copyright © 2017 Arm Ltd. All rights reserved.
// SPDX-License-Identifier: MIT
//
#pragma once
#include <armnn/Types.hpp>
#include <armnn/IRuntime.hpp>
#include <armnn/Deprecated.hpp>
#include <ISubgraphViewConverter.hpp>
#include <armnn/backends/IBackendContext.hpp>
#include <armnn/backends/IMemoryManager.hpp>
#include <armnn/backends/ITensorHandleFactory.hpp>
#include <armnn/backends/OptimizationViews.hpp>
#include <armnn/backends/SubgraphView.hpp>
#include <armnn/backends/profiling/IBackendProfiling.hpp>
#include <armnn/backends/profiling/IBackendProfilingContext.hpp>
#include <vector>
#include <memory>
namespace armnn
{
class IWorkloadFactory;
class IMemoryManager;
class ILayerSupport;
struct BackendVersion
{
uint32_t m_Major;
uint32_t m_Minor;
constexpr BackendVersion()
: m_Major(0)
, m_Minor(0)
{}
constexpr BackendVersion(uint32_t major, uint32_t minor)
: m_Major(major)
, m_Minor(minor)
{}
bool operator==(const BackendVersion& other) const
{
return this == &other ||
(this->m_Major == other.m_Major &&
this->m_Minor == other.m_Minor);
}
bool operator<=(const BackendVersion& other) const
{
return this->m_Major < other.m_Major ||
(this->m_Major == other.m_Major &&
this->m_Minor <= other.m_Minor);
}
bool operator>=(const BackendVersion& other) const
{
return this->m_Major > other.m_Major ||
(this->m_Major == other.m_Major &&
this->m_Minor >= other.m_Minor);
}
};
inline std::ostream& operator<<(std::ostream& os, const BackendVersion& backendVersion)
{
os << "[" << backendVersion.m_Major << "." << backendVersion.m_Minor << "]";
return os;
}
class IBackendInternal : public IBackend
{
protected:
/// Creation must be done through a specific
/// backend interface.
IBackendInternal() = default;
public:
/// Allow backends created by the factory function
/// to be destroyed through IBackendInternal.
~IBackendInternal() override = default;
using IWorkloadFactoryPtr = std::unique_ptr<IWorkloadFactory>;
using IBackendContextPtr = std::unique_ptr<IBackendContext>;
/// This is the bridge between backend and backend profiling we'll keep it in the backend namespace.
using IBackendProfilingContextPtr = std::shared_ptr<arm::pipe::IBackendProfilingContext>;
using IBackendProfilingPtr = std::unique_ptr<arm::pipe::IBackendProfiling>;
using ILayerSupportSharedPtr = std::shared_ptr<ILayerSupport>;
using IBackendSpecificModelContextPtr = std::shared_ptr<IBackendModelContext>;
using IMemoryManagerUniquePtr = std::unique_ptr<IMemoryManager>;
using IMemoryManagerSharedPtr = std::shared_ptr<IMemoryManager>;
virtual IMemoryManagerUniquePtr CreateMemoryManager() const;
virtual IWorkloadFactoryPtr CreateWorkloadFactory(
const IMemoryManagerSharedPtr& memoryManager = nullptr) const = 0;
virtual IWorkloadFactoryPtr CreateWorkloadFactory(
class TensorHandleFactoryRegistry& tensorHandleFactoryRegistry) const;
virtual IWorkloadFactoryPtr CreateWorkloadFactory(
const IMemoryManagerSharedPtr& memoryManager,
const ModelOptions& modelOptions) const;
virtual IWorkloadFactoryPtr CreateWorkloadFactory(
class TensorHandleFactoryRegistry& tensorHandleFactoryRegistry,
const ModelOptions& modelOptions) const;
virtual IWorkloadFactoryPtr CreateWorkloadFactory(
class TensorHandleFactoryRegistry& tensorHandleFactoryRegistry,
const ModelOptions& modelOptions,
MemorySourceFlags inputFlags,
MemorySourceFlags outputFlags) const;
/// Create the runtime context of the backend
///
/// Implementations may return a default-constructed IBackendContextPtr if
/// no context is needed at runtime.
/// Implementations must throw BackendUnavailableException if the backend
/// cannot be used (for example, necessary accelerator hardware is not present).
/// The default implementation always returns a default-constructed pointer.
virtual IBackendContextPtr CreateBackendContext(const IRuntime::CreationOptions&) const;
virtual IBackendSpecificModelContextPtr CreateBackendSpecificModelContext(const ModelOptions& modelOptions) const;
/// Create context specifically used for profiling interaction from backends.
virtual IBackendProfilingContextPtr CreateBackendProfilingContext(const IRuntime::CreationOptions& creationOptions,
IBackendProfilingPtr& backendProfiling);
virtual ILayerSupportSharedPtr GetLayerSupport() const = 0;
virtual ILayerSupportSharedPtr GetLayerSupport(const ModelOptions& modelOptions) const;
virtual OptimizationViews OptimizeSubgraphView(const SubgraphView& subgraph) const;
virtual OptimizationViews OptimizeSubgraphView(const SubgraphView& subgraph,
const ModelOptions& modelOptions) const;
bool SupportsTensorAllocatorAPI() const;
ITensorHandleFactory::FactoryId GetBackwardCompatibleFavoriteHandleFactory();
/// (Optional) Returns a vector of supported TensorHandleFactory ids in preference order.
virtual std::vector<ITensorHandleFactory::FactoryId> GetHandleFactoryPreferences() const;
/// (Optional) Register TensorHandleFactories
/// Either this method or CreateMemoryManager() and
/// IWorkloadFactory::CreateTensor()/IWorkloadFactory::CreateSubtensor() methods must be implemented.
virtual void RegisterTensorHandleFactories(class TensorHandleFactoryRegistry& /*registry*/) {}
/// (Optional) Register TensorHandleFactories
/// Either this method or CreateMemoryManager() and
/// IWorkloadFactory::CreateTensor()/IWorkloadFactory::CreateSubtensor() methods must be implemented.
virtual void RegisterTensorHandleFactories(class TensorHandleFactoryRegistry& registry,
MemorySourceFlags inputFlags,
MemorySourceFlags outputFlags);
/// Returns the version of the Backend API
static constexpr BackendVersion GetApiVersion() { return BackendVersion(1, 0); }
/// Returns a BackendCapability if the backend lists the capability
/// The BackendCapability must then be inspected to check whether or not that BackendCapability is supported
/// Otherwise returns an EmptyOptional if the BackendCapability is unlisted
virtual BackendCapabilities GetCapabilities() const
{
return BackendCapabilities("IBackendInternal NullCapabilities");
};
/// Returns true if backend support the capability false otherwise
ARMNN_DEPRECATED_MSG_REMOVAL_DATE("This function has been deprecated in favour of GetCapability", "22.05")
virtual bool HasCapability(BackendCapability /*capabilityClass*/) const { return false; }
/// Signals the backend to use a custom memory allocator provided by the user
///
/// \param allocator - a pointer to the provided ICustomAllocator to use with this backend
/// \param errMsg - Optional string variable to return error messages
/// \return - Returns true if switching to custom allocator was successful
virtual bool UseCustomMemoryAllocator(std::shared_ptr<ICustomAllocator> allocator,
armnn::Optional<std::string&> errMsg)
{
IgnoreUnused(allocator);
if (errMsg)
{
std::stringstream message;
message << "The backend " << GetId() << " doesn't support using a custom allocator. This error might"
" be related with the protected mode if the backend doesn't"
" fully support it.";
errMsg.value() = message.str();
}
return false;
}
/// Returns the default memory allocator for the backend
///
/// \return - Returns unique pointer to the Default Allocator of the Backend
virtual std::unique_ptr<ICustomAllocator> GetDefaultAllocator() const
{
throw armnn::Exception("GetDefaultAllocator: Function has not been implemented in backend.");
}
/// Returns the number of files cached if backend supports caching
///
/// \return - Returns 0 if backend does not support caching otherwise number of files cached
virtual unsigned int GetNumberOfCacheFiles() const { return 0; }
};
using IBackendInternalUniquePtr = std::unique_ptr<IBackendInternal>;
} // namespace armnn