| /* |
| * Copyright (c) 2024 Arm Limited. |
| * |
| * SPDX-License-Identifier: MIT |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a copy |
| * of this software and associated documentation files (the "Software"), to |
| * deal in the Software without restriction, including without limitation the |
| * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or |
| * sell copies of the Software, and to permit persons to whom the Software is |
| * furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included in all |
| * copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
| * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
| * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| * SOFTWARE. |
| */ |
| #include "helpers.h" |
| #include "tile_helpers.h" |
| |
| // The below defines the various reduce operations for our purposes. |
| // Where a corresponds to the existing value, and b the new value. |
| #define ADD_OP(a, b) ((a) + (b)) |
| #define SUB_OP(a, b) ((a) - (b)) |
| #define MAX_OP(a, b) fmax(a, b) |
| #define MIN_OP(a, b) fmin(a, b) |
| #define UPDATE_OP(a, b) (b) |
| |
| #ifdef SCATTER_MP1D_2D_MPND |
| |
| /** This kernel performs scatter operation |
| * |
| * @note Datatype should be given as a compile-time argument using -DDATA_TYPE=type. e.g. -DDATA_TYPE=short |
| * @note Number of indices should be given as a compile-time argument using -DNUM_INDICES, e.g. -DNUM_INDICES=3 |
| * @note Index length should be given as a compile-time argument using -DINDEX_LENGTH, e.g. -DINDEX_LENGTH=2 |
| * @note Outermost output shapes should be given as a compile-time argument using -DOUT_SHAPE_N_MINUS_X, where |
| * X must be 1,2,3,4,5, e.g. -DOUT_SHAPE_N_MINUS_1=3, ... |
| * @note Number of elements to copy in a row should be given as a compile-time argument using -DN0, e.g. -DN0=4 |
| * @note Number of partial elements at the edge to copy in a row should be given as a compile-time argument using |
| * -DPARTIAL_N0, e.g. -DPARTIAL_N0=2 |
| * @note Scatter function should be given as a compile-time argument using -DSCATTER_FUNCTION, e.g. -DSCATTER_FUNCTION=ADD |
| * @note If the kernel should skip reading the output tensor, -DSKIP_OUTPUT_READ option should be provided. |
| * @note Kernel name in uppercase letters should be provided as a compile-time argument, e.g. -DSCATTER_MP1D_2D_MPND |
| * |
| * @param[in] updates_ptr Pointer to the updates tensor. Data Types: F32 |
| * @param[in] updates_stride_x Stride of the updates tensor in X dimension (in bytes) |
| * @param[in] updates_step_x updates_stride_x * number of elements along X processed per workitem(in bytes) |
| * @param[in] updates_stride_y Stride of the updates tensor in Y dimension (in bytes) |
| * @param[in] updates_step_y updates_stride_y * number of elements along Y processed per workitem(in bytes) |
| * @param[in] updates_offset_first_element_in_bytes The offset of the first element in the updates tensor |
| * @param[in] indices_ptr Pointer to the indices tensor. Data Types: S32 |
| * @param[in] indices_stride_x Stride of the indices tensor in X dimension (in bytes) |
| * @param[in] indices_step_x indices_stride_x * number of elements along X processed per workitem(in bytes) |
| * @param[in] indices_stride_y Stride of the indices tensor in Y dimension (in bytes) |
| * @param[in] indices_step_y indices_stride_y * number of elements along Y processed per workitem(in bytes) |
| * @param[in] indices_offset_first_element_in_bytes The offset of the first element in the indices tensor |
| * @param[out] output_ptr Pointer to the destination tensor. Same as @p upt_ptr |
| * @param[in] output_stride_x Stride of the destination tensor in X dimension (in bytes) |
| * @param[in] output_step_x output_stride_x * number of elements along X processed per workitem(in bytes) |
| * @param[in] output_stride_y Stride of the destination tensor in Y dimension (in bytes) |
| * @param[in] output_step_y output_stride_y * number of elements along Y processed per workitem(in bytes) |
| * @param[in] output_offset_first_element_in_bytes The offset of the first element in the destination tensor |
| * @param[in] upt_block_stride Update tensor data block stride in bytes |
| * @param[in] out_block_stride Output tensor data block stride in bytes |
| */ |
| __kernel void scatter_mp1d_2d_mpnd( |
| IMAGE_DECLARATION(updates), |
| IMAGE_DECLARATION(indices), |
| IMAGE_DECLARATION(output), |
| int upt_block_stride, |
| int out_block_stride |
| ) |
| { |
| const int out_shape[5] = {OUT_SHAPE_N_MINUS_1, OUT_SHAPE_N_MINUS_2, OUT_SHAPE_N_MINUS_3, |
| OUT_SHAPE_N_MINUS_4, OUT_SHAPE_N_MINUS_5}; |
| |
| const int x = GET_SPATIAL_IDX(0, N0, PARTIAL_N0); // x-coordinate in the tensor |
| const int y = get_global_id(1); // collapsed y-coordinate (ignoring the outermost dimensions) |
| |
| const bool x_cond = (PARTIAL_N0 != 0 && get_global_id(0) == 0); |
| |
| uchar *ind_ptr_raw = indices_ptr + indices_offset_first_element_in_bytes; |
| const uchar *out_ptr_raw = output_ptr + output_offset_first_element_in_bytes |
| + x * sizeof(DATA_TYPE) + y * output_stride_y; |
| |
| const uchar *upt_ptr_raw = updates_ptr + updates_offset_first_element_in_bytes |
| + x * sizeof(DATA_TYPE) + y * updates_stride_y; |
| |
| for(int index_element = 0; index_element < NUM_INDICES; ++index_element) |
| { |
| const int *ind_ptr = (const int *) (ind_ptr_raw); |
| |
| // Out of bounds check |
| bool out_of_bounds = false; |
| LOOP_UNROLLING(int, i, 0, 1, INDEX_LENGTH, |
| { |
| if(ind_ptr[i] >= out_shape[i] || ind_ptr[i] < 0) |
| { |
| out_of_bounds = true; |
| } |
| }); |
| |
| ind_ptr_raw += indices_stride_y; |
| |
| if(out_of_bounds) |
| { |
| continue; |
| } |
| |
| // Index calculation |
| int index = 0; |
| LOOP_UNROLLING(int, i, 0, 1, INDEX_LENGTH, |
| { |
| index = index * out_shape[i] + ind_ptr[i]; |
| }); |
| |
| DATA_TYPE *out_ptr = (DATA_TYPE *) (out_ptr_raw + index * out_block_stride); |
| |
| const DATA_TYPE *upt_ptr = (const DATA_TYPE *) (upt_ptr_raw + index_element * upt_block_stride); |
| |
| VEC_DATA_TYPE(DATA_TYPE, N0) data_in0 = VLOAD(N0)(0, (__global DATA_TYPE *) upt_ptr); |
| |
| #ifdef SKIP_OUTPUT_READ |
| STORE_VECTOR_SELECT(data_in, DATA_TYPE, (__global DATA_TYPE *) out_ptr, N0, PARTIAL_N0, x_cond); |
| #else // ifdef SKIP_OUTPUT_READ |
| VEC_DATA_TYPE(DATA_TYPE, N0) data_out0 = VLOAD(N0)(0, (__global DATA_TYPE *) out_ptr); |
| data_out0 = SCATTER_FUNCTION(data_out0, data_in0); |
| |
| STORE_VECTOR_SELECT(data_out, DATA_TYPE, (__global DATA_TYPE *) out_ptr, N0, PARTIAL_N0, x_cond); |
| #endif // ifdef SKIP_OUTPUT_READ |
| } |
| } |
| |
| #endif // SCATTER_MP1D_2D_MPND |
| |
| #ifdef SCATTER1D_PARALLEL |
| |
| // NOTE : This code is non-deterministic and can only be excecuted with the "update" ScatterFunction |
| // This code is currently unusued as it requires changes to the existing test suite. |
| /** Performs the Scatter1D operation with multiple threads. |
| * Similar to @ref scatter1D() |
| */ |
| __kernel void scatter1D_parallel( |
| TENSOR4D_DECLARATION(updates), |
| TENSOR4D_DECLARATION(indices), |
| TENSOR4D_DECLARATION(output)) |
| { |
| // Currently 1D - only iterate through x dimension of indices. |
| const int px = get_global_id(0); |
| const int index_value = *(uchar*)(indices_ptr + indices_offset_first_element_in_bytes + (sizeof(int) * px)); |
| |
| if(index_value < OUT_SHAPE_X) |
| { |
| const DATA_TYPE update = *(DATA_TYPE *)(updates_ptr + updates_offset_first_element_in_bytes + (sizeof(DATA_TYPE) * px)); |
| __global uchar *out_addr = output_ptr + indices_offset_first_element_in_bytes + (sizeof(DATA_TYPE) * index_value); |
| *(__global DATA_TYPE *)(out_addr) = update; |
| } |
| } |
| |
| #endif // SCATTER1D_PARALLEL |