mirror of
https://github.com/ROCm/composable_kernel.git
synced 2026-05-01 12:11:19 +00:00
[CK_TILE] Update Stream-K Reduction Strategy Enum
## Motivation
Currently, Stream-K has 3 reduction options: 1) atomics, 2) The
reduction described in the Stream-K paper, and 3) a tree reduction. The
reduction strategy described in the original Stream-K paper has the
starting workgroup of each tile sequentially accumulating partial
results of other contributing workgroups in the tile, which requires a
linear number of steps. Hence, for clarity, this works updates the
naming of the `StreamKReductionStrategy` enum members to better describe
the existing reduction strategy options.
## Technical Details
Prior to this change, the enum is as follows:
```cpp
enum StreamKReductionStrategy : uint32_t
{
Atomic = 0u,
Reduction = 1u,
TreeReduction = 2u
};
```
But, the distinction between `Reduction` and `TreeReduction` is not very
clear and has some redundancy.
Hence, the updated enum is as follows:
```cpp
enum StreamKReductionStrategy : uint32_t
{
Atomic = 0u,
Linear = 1u,
Tree = 2u
};
```
All references to `StreamKReductionStrategy` were updated to reflect
this change.
## Test Plan
No new functionality was added, so no new tests were added; I just
validated existing tests and examples.
## Test Result
All tests passed locally.
## Submission Checklist
- [x] Look over the contributing guidelines at
https://github.com/ROCm/ROCm/blob/develop/CONTRIBUTING.md#pull-requests.
86 lines
3.5 KiB
C++
86 lines
3.5 KiB
C++
// Copyright (c) Advanced Micro Devices, Inc., or its affiliates.
|
|
// SPDX-License-Identifier: MIT
|
|
|
|
#pragma once
|
|
#include "ck_tile/core.hpp"
|
|
#include "ck_tile/host/kernel_launch.hpp"
|
|
#include "ck_tile/ops/epilogue.hpp"
|
|
#include "ck_tile/ops/gemm.hpp"
|
|
|
|
struct GemmConfigurationBase
|
|
{
|
|
static constexpr bool PAD_M = true;
|
|
static constexpr bool PAD_N = true;
|
|
static constexpr bool PAD_K = true;
|
|
|
|
static constexpr bool PERMUTE_A = false;
|
|
static constexpr bool PERMUTE_B = false;
|
|
|
|
static constexpr bool TRANSPOSE_C = false;
|
|
static constexpr bool USE_STRUCTURED_SPARSITY = false;
|
|
|
|
static constexpr int BLOCK_PER_CU = 1;
|
|
static constexpr auto SCHEDULER = ck_tile::GemmPipelineScheduler::Intrawave;
|
|
static constexpr ck_tile::index_t NUM_WAVE_GROUPS = 1;
|
|
static constexpr bool PRESHUFFLE = false;
|
|
static constexpr bool DOUBLE_SMEM_BUFFER = false;
|
|
};
|
|
|
|
template <typename PrecisionType, bool IsPersistent>
|
|
struct GemmConfigurationMemoryInterwave : public GemmConfigurationBase
|
|
{
|
|
static constexpr ck_tile::index_t M_TILE = 256;
|
|
static constexpr ck_tile::index_t N_TILE = 256;
|
|
static constexpr ck_tile::index_t K_TILE = 16;
|
|
|
|
static constexpr ck_tile::index_t M_WARP = 2;
|
|
static constexpr ck_tile::index_t N_WARP = 2;
|
|
static constexpr ck_tile::index_t K_WARP = 1;
|
|
|
|
static constexpr ck_tile::index_t M_WARP_TILE = 32;
|
|
static constexpr ck_tile::index_t N_WARP_TILE = 32;
|
|
static constexpr ck_tile::index_t K_WARP_TILE = sizeof(PrecisionType) == 2 ? 8 : 16;
|
|
|
|
static constexpr bool PERSISTENT = IsPersistent;
|
|
static constexpr auto SCHEDULER = ck_tile::GemmPipelineScheduler::Intrawave;
|
|
};
|
|
|
|
template <typename ADataType_, typename BDataType_ = ADataType_, typename CDataType_ = ADataType_>
|
|
struct StreamKGemmTypeConfiguration
|
|
{
|
|
using ADataType = ADataType_;
|
|
using BDataType = BDataType_;
|
|
using AccDataType = float;
|
|
using CDataType = CDataType_;
|
|
};
|
|
|
|
auto createArgs(int argc, char* argv[])
|
|
{
|
|
ck_tile::ArgParser arg_parser;
|
|
arg_parser.insert("m", "512", "m dimension")
|
|
.insert("n", "512", "n dimension")
|
|
.insert("k", "512", "k dimension")
|
|
.insert("a_layout", "R", "A tensor data layout - Row by default")
|
|
.insert("b_layout", "C", "B tensor data layout - Column by default")
|
|
.insert("c_layout", "R", "C tensor data layout - Row by default")
|
|
.insert("reduction_strategy",
|
|
"atomic",
|
|
"strategy for storing results in C tensor - atomic/linear")
|
|
.insert("persistent_dp",
|
|
"0",
|
|
"0. Non-persistent data-parallel section, 1 Fully persistent kernel.")
|
|
.insert("stride_a", "0", "Tensor A stride")
|
|
.insert("stride_b", "0", "Tensor B stride")
|
|
.insert("stride_c", "0", "Tensor C stride")
|
|
.insert("v", "2", "0. No validation, 1. Validation on CPU, 2. Validation on GPU")
|
|
.insert("prec", "fp16", "data type. fp16/bf16/fp8/bf8")
|
|
.insert("warmup", "50", "number of iterations before benchmarking the kernel")
|
|
.insert("repeat", "100", "number of iterations to benchmark the kernel")
|
|
.insert("timer", "gpu", "gpu:gpu timer, cpu:cpu timer")
|
|
.insert("init", "0", "0:random, 1:linear, 2:constant(1)")
|
|
.insert("flush_cache", "true", "flush cache before running the kernel, defaults to true");
|
|
|
|
bool result = arg_parser.parse(argc, argv);
|
|
return std::make_tuple(result, arg_parser);
|
|
}
|