mirror of
https://github.com/ROCm/composable_kernel.git
synced 2026-03-24 09:07:39 +00:00
* chore(copyright): update copyright header for codegen directory * chore(copyright): update copyright header for example directory
137 lines
4.8 KiB
C++
137 lines
4.8 KiB
C++
// Copyright (c) Advanced Micro Devices, Inc., or its affiliates.
|
|
// SPDX-License-Identifier: MIT
|
|
|
|
#pragma once
|
|
|
|
#include <cassert>
|
|
#include <vector>
|
|
|
|
#include "ck/ck.hpp"
|
|
#include "ck/utility/sequence.hpp"
|
|
#include "ck/utility/tuple.hpp"
|
|
#include "ck/utility/reduction_operator.hpp"
|
|
#include "ck/tensor_operation/gpu/device/impl/device_elementwise_dynamic_vector_dims_impl.hpp"
|
|
|
|
#include "batchnorm_common.hpp"
|
|
|
|
template <typename XDataType,
|
|
typename YDataType,
|
|
typename AccDataType,
|
|
typename ScaleDataType,
|
|
typename BiasDataType,
|
|
typename MeanVarDataType,
|
|
ck::index_t Rank,
|
|
ck::index_t NumBatchNormReduceDim,
|
|
bool fastest_dim_is_reduced = false>
|
|
int bnorm_infer(
|
|
bool time_kernel,
|
|
const std::array<int, NumBatchNormReduceDim> reduceDims,
|
|
const std::array<ck::index_t, Rank> xyLengths,
|
|
const std::array<ck::index_t, Rank> xStrides,
|
|
const std::array<ck::index_t, Rank> yStrides,
|
|
const std::array<ck::index_t, Rank - NumBatchNormReduceDim> bnScaleBiasMeanVarLengths,
|
|
const std::array<ck::index_t, Rank - NumBatchNormReduceDim> bnScaleStrides,
|
|
const std::array<ck::index_t, Rank - NumBatchNormReduceDim> bnBiasStrides,
|
|
const std::array<ck::index_t, Rank - NumBatchNormReduceDim> bnMeanVarStrides,
|
|
const void* p_x,
|
|
const void* p_scale,
|
|
const void* p_bias,
|
|
double epsilon,
|
|
const void* p_estimatedMean,
|
|
const void* p_estimatedVariance,
|
|
void* p_y)
|
|
{
|
|
(void)bnScaleBiasMeanVarLengths;
|
|
|
|
static_assert(NumBatchNormReduceDim < Rank,
|
|
"Invalid number of reduced dimensions for batchnorm!");
|
|
|
|
using DeviceNormalizeInstance = ck::tensor_operation::device::DeviceElementwiseImpl<
|
|
ck::Tuple<XDataType, AccDataType, AccDataType, AccDataType, AccDataType>, // x, mean,
|
|
// variance,
|
|
// scale,
|
|
// bias,
|
|
ck::Tuple<YDataType>, // y
|
|
NormalizeInInfer,
|
|
Rank,
|
|
64, // BlockSize
|
|
32, // MPerBlock
|
|
32, // NPerBlock
|
|
4, // MPerthread
|
|
4, // NPerthread
|
|
ck::Sequence<1, 0>, // ThreadClusterArrangeOrder
|
|
ck::Sequence<1, 1, 1, 1, 1>, // x, mean, variance, scale, bias
|
|
ck::Sequence<1>>; // scalarPerVector: y
|
|
|
|
auto invariantDims = get_invariant_dims<Rank, NumBatchNormReduceDim>(reduceDims);
|
|
std::array<ck::index_t, Rank> aligned_bnScaleStrides{0};
|
|
std::array<ck::index_t, Rank> aligned_bnBiasStrides{0};
|
|
std::array<ck::index_t, Rank> aligned_bnMeanVarStrides{0};
|
|
|
|
int i = 0;
|
|
for(auto dim : invariantDims)
|
|
{
|
|
assert(xyLengths[dim] == bnScaleBiasMeanVarLengths[i]);
|
|
|
|
aligned_bnScaleStrides[dim] = bnScaleStrides[i];
|
|
aligned_bnBiasStrides[dim] = bnBiasStrides[i];
|
|
aligned_bnMeanVarStrides[dim] = bnMeanVarStrides[i];
|
|
i++;
|
|
};
|
|
|
|
int32_t reduceLength = 1;
|
|
|
|
for(auto dim : reduceDims)
|
|
reduceLength *= xyLengths[dim];
|
|
|
|
int32_t invariantLength = 1;
|
|
|
|
for(auto dim : invariantDims)
|
|
invariantLength *= xyLengths[dim];
|
|
|
|
size_t total_length = static_cast<size_t>(invariantLength) * reduceLength;
|
|
|
|
float avg_time = 0.0f;
|
|
std::size_t num_bytes = 0;
|
|
|
|
auto dev_normalize = DeviceNormalizeInstance{};
|
|
|
|
auto argument_ptr1 = dev_normalize.MakeArgumentPointer(
|
|
xyLengths,
|
|
{xStrides,
|
|
aligned_bnMeanVarStrides,
|
|
aligned_bnMeanVarStrides,
|
|
aligned_bnScaleStrides,
|
|
aligned_bnBiasStrides},
|
|
{yStrides},
|
|
{p_x, p_estimatedMean, p_estimatedVariance, p_scale, p_bias},
|
|
{p_y},
|
|
NormalizeInInfer{epsilon});
|
|
|
|
if(!dev_normalize.IsSupportedArgument(argument_ptr1.get()))
|
|
{
|
|
std::cout << "The runtime parameters seems not supported by the Devic, exiting!"
|
|
<< std::endl;
|
|
|
|
return (-1);
|
|
};
|
|
|
|
auto invoker_ptr1 = dev_normalize.MakeInvokerPointer();
|
|
|
|
avg_time += invoker_ptr1->Run(argument_ptr1.get(), StreamConfig{nullptr, time_kernel});
|
|
|
|
num_bytes += total_length * sizeof(XDataType) +
|
|
invariantLength *
|
|
(sizeof(ScaleDataType) + sizeof(BiasDataType) + 2 * sizeof(MeanVarDataType)) +
|
|
total_length * sizeof(YDataType);
|
|
|
|
if(time_kernel)
|
|
{
|
|
float gb_per_sec = num_bytes / 1.E6 / avg_time;
|
|
|
|
std::cout << "Perf: " << avg_time << " ms, " << gb_per_sec << " GB/s" << std::endl;
|
|
};
|
|
|
|
return (0);
|
|
};
|