Enable padding blockscale for abquant (#3453)

* Enable padding blockscale for abquant

* run clang-format

* Reduce unnecessary testing

* remove cout
This commit is contained in:
kensclin
2025-12-25 01:12:40 +08:00
committed by GitHub
parent 1c3151963b
commit 7f68f3c4fa
8 changed files with 128 additions and 138 deletions

View File

@@ -412,32 +412,6 @@ struct QuantGemmKernel
return false;
}
if constexpr(kQuantType == QuantType::AQuantGrouped ||
kQuantType == QuantType::ABQuantGrouped)
{
if(kargs.QK_A % GemmPipeline::GetVectorSizeAQ() != 0)
{
if(ck_tile::EnvIsEnabled(CK_TILE_ENV(CK_TILE_LOGGING)))
{
CK_TILE_ERROR("K_A is not a multiple of vector load size for A tensor!");
}
return false;
}
}
if constexpr(kQuantType == QuantType::BQuantGrouped ||
kQuantType == QuantType::ABQuantGrouped)
{
if(kargs.QK_B % GemmPipeline::GetVectorSizeBQ() != 0)
{
if(ck_tile::EnvIsEnabled(CK_TILE_ENV(CK_TILE_LOGGING)))
{
CK_TILE_ERROR("K_B is not a multiple of vector load size for B tensor!");
}
return false;
}
}
if constexpr(std::is_same_v<ALayout, tensor_layout::gemm::RowMajor>)
{
if(kargs.K % (TilePartitioner::KPerBlock * kargs.k_batch) != 0 &&