blob: f19ad6dfc56c516a76101dcaa59b76e6d9de1dd1 [file] [log] [blame]
/*
* Copyright (c) 2022-2023 Arm Limited.
*
* SPDX-License-Identifier: MIT
*
* Permission is hereby granted, free of charge, to any person obtaining a copy
* of this software and associated documentation files (the "Software"), to
* deal in the Software without restriction, including without limitation the
* rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
* sell copies of the Software, and to permit persons to whom the Software is
* furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in all
* copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
#ifndef ARM_COMPUTE_DYNAMIC_FUSION_SKETCH_GPU_GPUWORKLOADSKETCH
#define ARM_COMPUTE_DYNAMIC_FUSION_SKETCH_GPU_GPUWORKLOADSKETCH
#include "arm_compute/core/TensorInfo.h"
#include "arm_compute/dynamic_fusion/sketch/gpu/GpuWorkloadContext.h"
#include <memory>
namespace arm_compute
{
namespace experimental
{
namespace dynamic_fusion
{
/** A descriptor of a workload of operators
*
* A "workload" is a basic unit of computation to schedule and perform. It contains one or more operators that can be "fused" together.
* Note that a workload may still contain multiple kernels.
*/
class GpuWorkloadSketch
{
public:
/** Global context used for the creation of a workload */
using Context = GpuWorkloadContext;
/** Internal opaque implementation */
class Implementation;
public:
/** Constructor
*
* @param[in] context Gpu context for the creation of a workload
*/
explicit GpuWorkloadSketch(GpuWorkloadContext *context);
/** Destructor */
~GpuWorkloadSketch();
/** Get the implementation */
Implementation &implementation();
/** Get the implementation */
const Implementation &implementation() const;
/** Get the gpu workload context of this sketch */
const GpuWorkloadContext *gpu_context() const;
/** Create a @ref TensorInfo associated with the workload sketch.
*
* @return TensorInfo Newly created tensor info
*/
template <typename... Args>
TensorInfo create_tensor_info(Args &&... args)
{
auto tensor_info = TensorInfo(std::forward<Args>(args)...);
tensor_info.set_id(allocate_new_tensor_id());
return tensor_info;
}
/** Create a @ref TensorInfo associated with the workload sketch by copying from an existing tensor info
* @note The newly copied tensor will have a different identity within the workload than the one copied from
* To copy the identity of @p tensor_info as well, use @ref TensorInfo 's copy constructors instead
*
* @param[in] tensor_info @ref ITensorInfo to copy from
*
* @return TensorInfo Newly created tensor info
*/
TensorInfo create_tensor_info(const ITensorInfo &tensor_info);
/** Create a default @ref TensorInfo associated with the workload sketch
* It is usually used by user input or output tensors
*
* @return TensorInfo Newly created tensor info
*/
TensorInfo create_tensor_info();
private:
ITensorInfo::Id allocate_new_tensor_id();
std::unique_ptr<Implementation> _impl; /**< Internal opaque implementation*/
};
} // namespace dynamic_fusion
} // namespace experimental
} // namespace arm_compute
#endif /* ARM_COMPUTE_DYNAMIC_FUSION_SKETCH_GPU_GPUWORKLOADSKETCH */