mirror of
https://github.com/ROCm/composable_kernel.git
synced 2026-05-14 10:09:41 +00:00
* convnd_fwd fp16 example
* update example
* update example
* update instance
* updating refernce conv
* update reference conv
* update conv fwd profiler
* update conv 1d and 3d instance
* update include path
* clean
* update profiler for conv bwd data and weight
* update conv bwd weight
* clean
* update conv example
* update profiler for conv bwd weight
* update ckprofiler for conv bwd data
* fix reference conv bwd data bug; update conv bwd data test
* update examples
* fix initialization issue
* update test for conv fwd
* clean
* clean
* remove test case too sensitive to error threshhold
* fix test
* clean
* fix build
* adding conv multiple d
* adding conv multiple D
* add matrix padder
* add gemm padding to convnd
* adding group conv
* update gemm multi-d
* refactor
* refactor
* refactor
* clean
* clean
* refactor
* refactor
* reorg
* add ds
* add bias
* clean
* add G
* adding group
* adding group
* adding group
* update Tensor
* clean
* update example
* update DeviceGemmMultipleD_Xdl_CShuffle
* update conv bwd-data and bwd-weight
* upate contraction example
* update gemm and batch gemm with e permute
* fix example build
* instance for grouped conv1d
* update example
* adding group conv instance
* update gemm bilinear instance
* update gemm+add+add+fastgelu instance
* update profiler
* update profiler
* update test
* update test and client example
* clean
* add grouped conv into profiler
* update profiler
* clean
* add test grouped conv, update all conv test to gtest
* update test
[ROCm/composable_kernel commit: 500fa99512]
242 lines
13 KiB
C++
242 lines
13 KiB
C++
// SPDX-License-Identifier: MIT
|
|
// Copyright (c) 2018-2022, Advanced Micro Devices, Inc. All rights reserved.
|
|
|
|
#include <cstdlib>
|
|
#include <iostream>
|
|
#include <initializer_list>
|
|
#include <vector>
|
|
#include <gtest/gtest.h>
|
|
|
|
#include "profiler/include/profile_conv_bwd_data_impl.hpp"
|
|
|
|
class TestConvndBwdData : public ::testing::Test
|
|
{
|
|
protected:
|
|
std::vector<ck::utils::conv::ConvParam> conv_params;
|
|
};
|
|
|
|
// 1d
|
|
TEST_F(TestConvndBwdData, Conv1dBwdData)
|
|
{
|
|
conv_params.clear();
|
|
conv_params.push_back({1, 1, 128, 128, 256, {1}, {14}, {2}, {1}, {0}, {0}});
|
|
conv_params.push_back({1, 1, 128, 128, 256, {3}, {28}, {1}, {1}, {1}, {1}});
|
|
conv_params.push_back({1, 1, 128, 128, 256, {1}, {3}, {1}, {1}, {0}, {0}});
|
|
|
|
for(auto& param : conv_params)
|
|
{
|
|
bool pass;
|
|
|
|
// fp32
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<1,
|
|
ck::tensor_layout::convolution::NWC,
|
|
ck::tensor_layout::convolution::KXC,
|
|
ck::tensor_layout::convolution::NWK,
|
|
float,
|
|
float,
|
|
float>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// fp16
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<1,
|
|
ck::tensor_layout::convolution::NWC,
|
|
ck::tensor_layout::convolution::KXC,
|
|
ck::tensor_layout::convolution::NWK,
|
|
ck::half_t,
|
|
ck::half_t,
|
|
ck::half_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// bf16
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<1,
|
|
ck::tensor_layout::convolution::NWC,
|
|
ck::tensor_layout::convolution::KXC,
|
|
ck::tensor_layout::convolution::NWK,
|
|
ck::bhalf_t,
|
|
ck::bhalf_t,
|
|
ck::bhalf_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// int8
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<1,
|
|
ck::tensor_layout::convolution::NWC,
|
|
ck::tensor_layout::convolution::KXC,
|
|
ck::tensor_layout::convolution::NWK,
|
|
int8_t,
|
|
int8_t,
|
|
int8_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
}
|
|
}
|
|
|
|
// 2d
|
|
TEST_F(TestConvndBwdData, Conv2dBwdData)
|
|
{
|
|
conv_params.clear();
|
|
conv_params.push_back({2, 1, 128, 128, 256, {1, 1}, {7, 7}, {2, 2}, {1, 1}, {0, 0}, {0, 0}});
|
|
conv_params.push_back({2, 1, 128, 128, 256, {3, 3}, {14, 14}, {1, 1}, {1, 1}, {1, 1}, {1, 1}});
|
|
conv_params.push_back({2, 1, 128, 128, 256, {1, 1}, {3, 3}, {1, 1}, {1, 1}, {0, 0}, {0, 0}});
|
|
|
|
for(auto& param : conv_params)
|
|
{
|
|
bool pass;
|
|
|
|
// fp32
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<2,
|
|
ck::tensor_layout::convolution::NHWC,
|
|
ck::tensor_layout::convolution::KYXC,
|
|
ck::tensor_layout::convolution::NHWK,
|
|
float,
|
|
float,
|
|
float>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// fp16
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<2,
|
|
ck::tensor_layout::convolution::NHWC,
|
|
ck::tensor_layout::convolution::KYXC,
|
|
ck::tensor_layout::convolution::NHWK,
|
|
ck::half_t,
|
|
ck::half_t,
|
|
ck::half_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// bf16
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<2,
|
|
ck::tensor_layout::convolution::NHWC,
|
|
ck::tensor_layout::convolution::KYXC,
|
|
ck::tensor_layout::convolution::NHWK,
|
|
ck::bhalf_t,
|
|
ck::bhalf_t,
|
|
ck::bhalf_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// int8
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<2,
|
|
ck::tensor_layout::convolution::NHWC,
|
|
ck::tensor_layout::convolution::KYXC,
|
|
ck::tensor_layout::convolution::NHWK,
|
|
int8_t,
|
|
int8_t,
|
|
int8_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
}
|
|
}
|
|
|
|
// 3d
|
|
TEST_F(TestConvndBwdData, Conv3dBwdData)
|
|
{
|
|
conv_params.clear();
|
|
conv_params.push_back(
|
|
{3, 1, 128, 128, 256, {1, 1, 1}, {7, 7, 7}, {2, 2, 2}, {1, 1, 1}, {0, 0, 0}, {0, 0, 0}});
|
|
conv_params.push_back(
|
|
{3, 1, 128, 128, 256, {3, 3, 3}, {14, 14, 3}, {1, 1, 1}, {1, 1, 1}, {1, 1, 1}, {1, 1, 1}});
|
|
conv_params.push_back(
|
|
{3, 1, 128, 128, 256, {1, 1, 1}, {3, 3, 3}, {1, 1, 1}, {1, 1, 1}, {0, 0, 0}, {0, 0, 0}});
|
|
|
|
for(auto& param : conv_params)
|
|
{
|
|
bool pass;
|
|
|
|
// fp32
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<3,
|
|
ck::tensor_layout::convolution::NDHWC,
|
|
ck::tensor_layout::convolution::KZYXC,
|
|
ck::tensor_layout::convolution::NDHWK,
|
|
float,
|
|
float,
|
|
float>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// fp16
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<3,
|
|
ck::tensor_layout::convolution::NDHWC,
|
|
ck::tensor_layout::convolution::KZYXC,
|
|
ck::tensor_layout::convolution::NDHWK,
|
|
ck::half_t,
|
|
ck::half_t,
|
|
ck::half_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// bf16
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<3,
|
|
ck::tensor_layout::convolution::NDHWC,
|
|
ck::tensor_layout::convolution::KZYXC,
|
|
ck::tensor_layout::convolution::NDHWK,
|
|
ck::bhalf_t,
|
|
ck::bhalf_t,
|
|
ck::bhalf_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
|
|
// int8
|
|
pass = ck::profiler::profile_conv_bwd_data_impl<3,
|
|
ck::tensor_layout::convolution::NDHWC,
|
|
ck::tensor_layout::convolution::KZYXC,
|
|
ck::tensor_layout::convolution::NDHWK,
|
|
int8_t,
|
|
int8_t,
|
|
int8_t>(true, // do_verification
|
|
1, // init_method
|
|
false, // do_log
|
|
false, // time_kernel
|
|
param);
|
|
|
|
EXPECT_TRUE(pass);
|
|
}
|
|
}
|