// Copyright (c) Advanced Micro Devices, Inc., or its affiliates. // SPDX-License-Identifier: MIT #include #include #include "ck/ck.hpp" #include "ck/tensor_operation/gpu/element/combined_element_wise_operation.hpp" #include "ck/tensor_operation/gpu/device/impl/device_elementwise_dynamic_vector_dims_impl.hpp" #include "ck/library/reference_tensor_operation/cpu/reference_elementwise.hpp" #include "ck/library/utility/algorithm.hpp" #include "ck/library/utility/check_err.hpp" #include "ck/library/utility/device_memory.hpp" #include "ck/library/utility/host_tensor.hpp" #include "ck/library/utility/host_tensor_generator.hpp" using ::ck::DeviceMem; using ::ck::HostTensorDescriptor; using ::ck::Tensor; using F16 = ck::half_t; using F32 = float; using ADataType = F16; using BDataType = F16; using NchwLayout = ck::tensor_layout::convolution::NCHW; using NhwcLayout = ck::tensor_layout::convolution::NHWC; using UnaryScale = ck::tensor_operation::element_wise::Scale; using UnarySquare = ck::tensor_operation::element_wise::UnarySquare; using UnaryScaleSquare = ck::tensor_operation::element_wise::UnaryCombinedOp; using BinaryAdd = ck::tensor_operation::element_wise::Add; // B = alpha * A0 * A0 + beta * A1 * A1 using BinaryAddUnaryScaleSquare = ck::tensor_operation::element_wise:: BinaryWithUnaryCombinedOp; using DeviceElementwisePermuteInstance = ck::tensor_operation::device::DeviceElementwiseImpl< ck::Tuple, // InDataTypeTuple ck::Tuple, // OutDataTypeTuple BinaryAddUnaryScaleSquare, // ElementwiseOp 4, // NumDim 256, // BlockSize 128, // M0PerBlock 128, // M1PerBlock 8, // M0PerThread 8, // M1PerThread ck::Sequence<1, 0>, // ThreadClusterArrangeOrder ck::Sequence<8, 8>, // InScalarPerVectorSeq ck::Sequence<8>>; // OutScalarPerVectorSeq int main(int argc, char* argv[]) { bool do_verification = true; bool time_kernel = false; std::vector nchw = {16, 128, 32, 64}; if(argc == 1) { // use default } else if(argc == 3) { do_verification = std::stoi(argv[1]); time_kernel = std::stoi(argv[2]); } else if(argc == 7) { do_verification = std::stoi(argv[1]); time_kernel = std::stoi(argv[2]); nchw[0] = std::stoi(argv[3]); nchw[1] = std::stoi(argv[4]); nchw[2] = std::stoi(argv[5]); nchw[3] = std::stoi(argv[6]); } else { printf("arg1: verification (0=no, 1=yes)\n"); printf("arg2: time kernel (0=no, 1=yes)\n"); printf("arg3-6: N, C, H, W (default 16, 128, 32, 64)\n"); exit(1); } std::array ab_lengths; std::array ab_strides = {static_cast(nchw[1] * nchw[2] * nchw[3]), static_cast(nchw[2] * nchw[3]), static_cast(nchw[3]), 1}; ck::ranges::copy(nchw, ab_lengths.begin()); std::array, 2> as = {Tensor(ab_lengths, ab_strides, NchwLayout{}), Tensor(ab_lengths, ab_strides, NchwLayout{})}; Tensor& a0 = as[0]; Tensor& a1 = as[1]; Tensor b(ab_lengths, ab_strides, NchwLayout{}); float alpha = 3.f; float beta = 2.f; a0.GenerateTensorValue(GeneratorTensor_3{0.0, 1.0}); a1.GenerateTensorValue(GeneratorTensor_3{0.0, 1.0}); DeviceMem a0_device_buf(sizeof(ADataType) * a0.mDesc.GetElementSpaceSize()); DeviceMem a1_device_buf(sizeof(ADataType) * a1.mDesc.GetElementSpaceSize()); DeviceMem b_device_buf(sizeof(BDataType) * b.mDesc.GetElementSpaceSize()); a0_device_buf.ToDevice(a0.mData.data()); a1_device_buf.ToDevice(a1.mData.data()); std::array inputs = {a0_device_buf.GetDeviceBuffer(), a1_device_buf.GetDeviceBuffer()}; std::array output = {b_device_buf.GetDeviceBuffer()}; auto broadcastPermute = DeviceElementwisePermuteInstance{}; auto unary_scale_op_a0 = UnaryScaleSquare{UnarySquare{}, UnaryScale{alpha}}; auto unary_scale_op_a1 = UnaryScaleSquare{UnarySquare{}, UnaryScale{beta}}; auto argument = broadcastPermute.MakeArgumentPointer( ab_lengths, {ab_strides, ab_strides}, {ab_strides}, inputs, output, BinaryAddUnaryScaleSquare{BinaryAdd{}, unary_scale_op_a0, unary_scale_op_a1}); if(!broadcastPermute.IsSupportedArgument(argument.get())) { throw std::runtime_error( "The runtime parameters seems not supported by the device instance, exiting!"); }; std::cout << "A0 (nchw): " << a0.mDesc << std::endl; std::cout << "A1 (nchw): " << a1.mDesc << std::endl; std::cout << "B (nchw): " << b.mDesc << std::endl; auto broadcastPermute_invoker_ptr = broadcastPermute.MakeInvokerPointer(); float ave_time = broadcastPermute_invoker_ptr->Run(argument.get(), StreamConfig{nullptr, time_kernel}); std::size_t flop = std::size_t(5) * nchw[0] * nchw[1] * nchw[2] * nchw[3]; std::size_t num_btype = sizeof(ADataType) * (nchw[0] * nchw[1] * nchw[2] * nchw[3]) + sizeof(BDataType) * (nchw[0] * nchw[1] * nchw[2] * nchw[3]); float tflops = static_cast(flop) / 1.E9 / ave_time; float gb_per_sec = num_btype / 1.E6 / ave_time; std::cout << "Perf: " << ave_time << " ms, " << tflops << " TFlops, " << gb_per_sec << " GB/s" << std::endl; bool pass = true; if(do_verification) { Tensor host_b(ab_lengths, ab_strides, NchwLayout{}); using ReferenceElementwiseInstance = ck::tensor_operation::host:: ReferenceElementwise<2, ADataType, BDataType, BinaryAddUnaryScaleSquare>; auto ref_elementwise = ReferenceElementwiseInstance{}; auto ref_invoker = ref_elementwise.MakeInvoker(); auto ref_argument = ref_elementwise.MakeArgument( as, host_b, BinaryAddUnaryScaleSquare{BinaryAdd{}, unary_scale_op_a0, unary_scale_op_a1}); ref_invoker.Run(ref_argument); b_device_buf.FromDevice(b.mData.data()); pass &= ck::utils::check_err(b.mData, host_b.mData, "Error: Incorrect results b", 1e-3, 1e-3); } return pass ? 0 : 1; }