// Copyright (c) Advanced Micro Devices, Inc., or its affiliates. // SPDX-License-Identifier: MIT #include #include #include "ck/ck.hpp" #include "ck/tensor_operation/gpu/element/combined_element_wise_operation.hpp" #include "ck/tensor_operation/gpu/device/impl/device_elementwise_dynamic_vector_dims_impl.hpp" #include "ck/library/reference_tensor_operation/cpu/reference_elementwise.hpp" #include "ck/library/utility/algorithm.hpp" #include "ck/library/utility/check_err.hpp" #include "ck/library/utility/device_memory.hpp" #include "ck/library/utility/host_tensor.hpp" #include "ck/library/utility/host_tensor_generator.hpp" using ::ck::DeviceMem; using ::ck::HostTensorDescriptor; using ::ck::Tensor; using F16 = ck::half_t; using F32 = float; using ADataType = F16; using BDataType = F16; using NchwLayout = ck::tensor_layout::convolution::NCHW; using NhwcLayout = ck::tensor_layout::convolution::NHWC; using UnaryScale = ck::tensor_operation::element_wise::Scale; using UnarySquare = ck::tensor_operation::element_wise::UnarySquare; using UnaryScaleSquare = ck::tensor_operation::element_wise::UnaryCombinedOp; using BinaryAdd = ck::tensor_operation::element_wise::Add; // B = alpha * A0 * A0 + beta * A1 * A1 + gamma * A2 * A2 using TrinaryAddUnaryScaleSquare = ck::tensor_operation::element_wise::TrinaryWithUnaryCombinedOp; using DeviceElementwisePermuteInstance = ck::tensor_operation::device::DeviceElementwiseImpl< ck::Tuple, // InDataTypeTuple ck::Tuple, // OutDataTypeTuple TrinaryAddUnaryScaleSquare, // ElementwiseOp 4, // NumDim 256, // BlockSize 128, // M0PerBlock 128, // M1PerBlock 8, // M0PerThread 8, // M1PerThread ck::Sequence<1, 0>, // ThreadClusterArrangeOrder ck::Sequence<8, 8, 8>, // InScalarPerVectorSeq ck::Sequence<8>>; // OutScalarPerVectorSeq int main(int argc, char* argv[]) { bool do_verification = true; bool time_kernel = false; if(argc == 1) { // use default } else if(argc == 3) { do_verification = std::stoi(argv[1]); time_kernel = std::stoi(argv[2]); } else { printf("arg1: verification (0=no, 1=yes)\n"); printf("arg2: time kernel (0=no, 1=yes)\n"); exit(0); } std::vector nchw = {16, 128, 32, 64}; std::array ab_lengths; std::array ab_strides = {static_cast(nchw[1] * nchw[2] * nchw[3]), static_cast(nchw[2] * nchw[3]), static_cast(nchw[3]), 1}; ck::ranges::copy(nchw, ab_lengths.begin()); std::array, 3> as = {Tensor(ab_lengths, ab_strides, NchwLayout{}), Tensor(ab_lengths, ab_strides, NchwLayout{}), Tensor(ab_lengths, ab_strides, NchwLayout{})}; Tensor& a0 = as[0]; Tensor& a1 = as[1]; Tensor& a2 = as[2]; Tensor b(ab_lengths, ab_strides, NchwLayout{}); float alpha = 3.f; float beta = 2.f; float gamma = 4.f; a0.GenerateTensorValue(GeneratorTensor_3{0.0, 1.0}); a1.GenerateTensorValue(GeneratorTensor_3{0.0, 1.0}); a2.GenerateTensorValue(GeneratorTensor_3{0.0, 1.0}); DeviceMem a0_device_buf(sizeof(ADataType) * a0.mDesc.GetElementSpaceSize()); DeviceMem a1_device_buf(sizeof(ADataType) * a1.mDesc.GetElementSpaceSize()); DeviceMem a2_device_buf(sizeof(ADataType) * a2.mDesc.GetElementSpaceSize()); DeviceMem b_device_buf(sizeof(BDataType) * b.mDesc.GetElementSpaceSize()); a0_device_buf.ToDevice(a0.mData.data()); a1_device_buf.ToDevice(a1.mData.data()); a2_device_buf.ToDevice(a2.mData.data()); std::array inputs = {a0_device_buf.GetDeviceBuffer(), a1_device_buf.GetDeviceBuffer(), a2_device_buf.GetDeviceBuffer()}; std::array output = {b_device_buf.GetDeviceBuffer()}; auto broadcastPermute = DeviceElementwisePermuteInstance{}; auto unary_scale_op_a0 = UnaryScaleSquare{UnarySquare{}, UnaryScale{alpha}}; auto unary_scale_op_a1 = UnaryScaleSquare{UnarySquare{}, UnaryScale{beta}}; auto unary_scale_op_a2 = UnaryScaleSquare{UnarySquare{}, UnaryScale{gamma}}; auto argument = broadcastPermute.MakeArgumentPointer( ab_lengths, {ab_strides, ab_strides, ab_strides}, {ab_strides}, inputs, output, TrinaryAddUnaryScaleSquare{ BinaryAdd{}, BinaryAdd{}, unary_scale_op_a0, unary_scale_op_a1, unary_scale_op_a2}); if(!broadcastPermute.IsSupportedArgument(argument.get())) { throw std::runtime_error( "The runtime parameters seems not supported by the device instance, exiting!"); }; std::cout << "A0 (nchw): " << a0.mDesc << std::endl; std::cout << "A1 (nchw): " << a1.mDesc << std::endl; std::cout << "A2 (nchw): " << a2.mDesc << std::endl; std::cout << "B (nchw): " << b.mDesc << std::endl; auto broadcastPermute_invoker_ptr = broadcastPermute.MakeInvokerPointer(); float ave_time = broadcastPermute_invoker_ptr->Run(argument.get(), StreamConfig{nullptr, time_kernel}); std::size_t flop = std::size_t(5) * nchw[0] * nchw[1] * nchw[2] * nchw[3]; std::size_t num_btype = sizeof(ADataType) * (nchw[0] * nchw[1] * nchw[2] * nchw[3]) + sizeof(BDataType) * (nchw[0] * nchw[1] * nchw[2] * nchw[3]); float tflops = static_cast(flop) / 1.E9 / ave_time; float gb_per_sec = num_btype / 1.E6 / ave_time; std::cout << "Perf: " << ave_time << " ms, " << tflops << " TFlops, " << gb_per_sec << " GB/s" << std::endl; bool pass = true; if(do_verification) { Tensor host_b(ab_lengths, ab_strides, NchwLayout{}); using ReferenceElementwiseInstance = ck::tensor_operation::host:: ReferenceElementwise<3, ADataType, BDataType, TrinaryAddUnaryScaleSquare>; auto ref_elementwise = ReferenceElementwiseInstance{}; auto ref_invoker = ref_elementwise.MakeInvoker(); auto ref_argument = ref_elementwise.MakeArgument( as, host_b, TrinaryAddUnaryScaleSquare{ BinaryAdd{}, BinaryAdd{}, unary_scale_op_a0, unary_scale_op_a1, unary_scale_op_a2}); ref_invoker.Run(ref_argument); const double threshold = std::pow(2, -10) * 2; b_device_buf.FromDevice(b.mData.data()); pass &= ck::utils::check_err( b.mData, host_b.mData, "Error: Incorrect results b", threshold, threshold); } return pass ? 0 : 1; }