// SPDX-License-Identifier: MIT // Copyright (c) 2024-2025, Advanced Micro Devices, Inc. All rights reserved. #include #include #include #include #include #include #include "ck_tile/host.hpp" #include "gemm_utils.hpp" template void try_run(ck_tile::TailNumber tn) { if constexpr(Pipeline::PrefetchStages > static_cast(TN)) { if(tn == TN) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } } } template float gemm_calc(const ck_tile::GemmHostArgs& args, const ck_tile::stream_config& s) { using GemmShape = ck_tile::TileGemmShape< ck_tile::sequence, ck_tile::sequence, ck_tile:: sequence, GemmConfig::PermuteA, GemmConfig::PermuteB>; using TilePartitioner = ck_tile::GemmSpatiallyLocalTilePartitioner; using Traits = ck_tile::TileGemmTraits; using GemmUniversalTraits = ck_tile::TileGemmUniversalTraits; using GemmPipelineProblem = ck_tile::GemmPipelineProblem; using BaseGemmPipeline = UNIVERSAL_GEMM_PIPELINE; const ck_tile::index_t k_grain = args.k_batch * GemmConfig::K_Tile; const ck_tile::index_t K_split = (args.K + k_grain - 1) / k_grain * GemmConfig::K_Tile; const ck_tile::index_t num_loop = TilePartitioner::GetLoopNum(K_split); const bool has_hot_loop = BaseGemmPipeline::BlockHasHotloop(num_loop); const ck_tile::TailNumber tail_num = BaseGemmPipeline::GetBlockLoopTailNum(num_loop); float ave_time{0}; const auto Run = [&](const auto has_hot_loop_, const auto tail_number_, const auto memory_operation_) { constexpr bool has_hot_loop_v = has_hot_loop_.value; constexpr auto tail_number_v = tail_number_.value; constexpr auto scheduler = GEMM_PIPELINE_SCHEDULER; constexpr auto memory_operation = memory_operation_.value; using UniversalGemmProblem = ck_tile::UniversalGemmPipelineProblem; using GemmPipeline = GEMM_PIPELINE; using GemmEpilogue = ck_tile::CShuffleEpilogue< ck_tile::CShuffleEpilogueProblem>; using Kernel = ck_tile::GemmKernel; auto kargs = Kernel::MakeKernelArgs(args); const dim3 grids = Kernel::GridSize(args.M, args.N, args.k_batch); constexpr dim3 blocks = Kernel::BlockSize(); if(!Kernel::IsSupportedArgument(kargs)) { throw std::runtime_error("Wrong! Arguments not supported! Skipping gemm!\n"); } if(s.log_level_ > 0) { std::cout << "Launching kernel with args:" << " grid: {" << grids.x << ", " << grids.y << ", " << grids.z << "}" << ", blocks: {" << blocks.x << ", " << blocks.y << ", " << blocks.z << "}" << std::endl; } ave_time = ck_tile::launch_kernel(s, ck_tile::make_kernel( Kernel{}, grids, blocks, 0, kargs)); return ave_time; }; const auto RunSplitk = [&](const auto has_hot_loop_, const auto tail_number_) { if(args.k_batch == 1) { Run(has_hot_loop_, tail_number_, ck_tile::integral_constant{}); } else { Run(has_hot_loop_, tail_number_, ck_tile::integral_constant{}); } }; if(has_hot_loop) { #if(CK_TILE_PIPELINE_DEFAULT == CK_TILE_PIPELINE_COMPUTE_V3) if(tail_num == ck_tile::TailNumber::Full) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } else if(tail_num == ck_tile::TailNumber::Odd) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } else if(tail_num == ck_tile::TailNumber::Even) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } else { std::ostringstream err; err << "For compute pipeline tail number should always be Full, but have \"" << tail_num << "\" which is not supported! PrefetchStages: " << BaseGemmPipeline::PrefetchStages << "\n File: " << __FILE__ << ":" << __LINE__ << ", in function: " << __func__; throw std::runtime_error(err.str()); } #elif(CK_TILE_PIPELINE_DEFAULT == CK_TILE_PIPELINE_MEMORY) if(tail_num == ck_tile::TailNumber::One) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } else if(tail_num == ck_tile::TailNumber::Full) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } auto check_tail = [&](auto... TNs) { (try_run(tail_num), ...); }; check_tail(ck_tile::integral_constant{}, ck_tile::integral_constant{}, ck_tile::integral_constant{}, ck_tile::integral_constant{}, ck_tile::integral_constant{}, ck_tile::integral_constant{}); #elif(CK_TILE_PIPELINE_DEFAULT == CK_TILE_PIPELINE_COMPUTE_V4) if(tail_num == ck_tile::TailNumber::Three) { RunSplitk( ck_tile::bool_constant{}, ck_tile::integral_constant{}); } else { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } #endif } else { if(tail_num == ck_tile::TailNumber::Full) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } else if(tail_num == ck_tile::TailNumber::Odd) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } else if(tail_num == ck_tile::TailNumber::Even) { RunSplitk(ck_tile::bool_constant{}, ck_tile::integral_constant{}); } else { std::ostringstream err; err << "Num K loop must be larger than number of prefetech stages." << "\n PrefetchStages: " << BaseGemmPipeline::PrefetchStages << "\n File: " << __FILE__ << ":" << __LINE__ << ", in function: " << __func__; throw std::runtime_error(err.str()); } } return ave_time; } #include "run_gemm_example.inc" template int run_gemm_example_prec_type(std::string a_layout, std::string b_layout, int argc, char* argv[]) { using Row = ck_tile::tensor_layout::gemm::RowMajor; using Col = ck_tile::tensor_layout::gemm::ColumnMajor; if constexpr(std::is_same_v) { if(a_layout == "R" && b_layout == "C") { return run_gemm_example_with_layouts( argc, argv, Row{}, Col{}, Row{}); } else if(a_layout == "C" && b_layout == "C") { return run_gemm_example_with_layouts( argc, argv, Col{}, Col{}, Row{}); } else { throw std::runtime_error("Unsupported memory layout for the input matrices when " "BPrecType is ck_tile::pk_int4_t!"); } } else { if(a_layout == "R" && b_layout == "R") { return run_gemm_example_with_layouts( argc, argv, Row{}, Row{}, Row{}); } else if(a_layout == "R" && b_layout == "C") { return run_gemm_example_with_layouts( argc, argv, Row{}, Col{}, Row{}); } else if(a_layout == "C" && b_layout == "R") { return run_gemm_example_with_layouts( argc, argv, Col{}, Row{}, Row{}); } else if(a_layout == "C" && b_layout == "C") { return run_gemm_example_with_layouts( argc, argv, Col{}, Col{}, Row{}); } else { throw std::runtime_error("Unsupported memory layout for the input matrices!"); } } } int run_gemm_example(int argc, char* argv[]) { auto [result, arg_parser] = create_args(argc, argv); if(!result) return -1; std::string data_type = arg_parser.get_str("prec"); std::string a_layout = arg_parser.get_str("a_layout"); std::string b_layout = arg_parser.get_str("b_layout"); if(data_type == "fp16") { return run_gemm_example_prec_type(a_layout, b_layout, argc, argv); } else if(data_type == "bf16") { return run_gemm_example_prec_type(a_layout, b_layout, argc, argv); } else if(data_type == "fp8") { return run_gemm_example_prec_type( a_layout, b_layout, argc, argv); } else if(data_type == "bf8") { return run_gemm_example_prec_type( a_layout, b_layout, argc, argv); } #if(CK_TILE_PIPELINE_DEFAULT == CK_TILE_PIPELINE_COMPUTE_V3) else if(data_type == "pk_int4_t") { // TODO: Add support for bhalf_t ADataType return run_gemm_example_prec_type( a_layout, b_layout, argc, argv); } #endif else { throw std::runtime_error("Unsupported data type for this operation !!!"); } } int main(int argc, char* argv[]) { try { run_gemm_example(argc, argv); } catch(const std::runtime_error& e) { std::cerr << "Caught runtime error: " << e.what() << '\n'; // Return a non-zero code to indicate failure return EXIT_FAILURE; } return EXIT_SUCCESS; }