Create operation support matrix for CK Tile Engine Introduce operation support matrix for CK Tile kernels detailing data types, layouts, and GPU targets. ## Motivation The tile engine currently supports a subset of CK Tile operations, but there is no in-repo reference that maps which operations, data types, layouts, and GPU targets are covered by the tile engine versus only available through hand-written examples or tests. This makes it difficult for developers to know what the tile engine already handles, what requires manual integration, and where coverage gaps exist. This PR introduces an operation support matrix as a markdown file in tile_engine/, intended to be maintained as a living document alongside the code. Because it lives in the repository rather than an external wiki or PDF, it can be reviewed and updated in the same pull requests that add or extend tile engine operations, keeping it accurate as coverage evolves. ## Technical Details Documentation only change. ## Test Plan N/A ## Test Result N/A ## Submission Checklist - [x] Look over the contributing guidelines at https://github.com/ROCm/ROCm/blob/develop/CONTRIBUTING.md#pull-requests.
7.2 KiB
CK Tile operation support by data type, layout, and GPU target:
| Op | CK Tile Kernel | fp16 | fp8 | bf16 | bf8 | int8 | fp4 | fp6 | rcr | rrr | ccr | crr | 90a | 942 | 950 | 1201 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| GEMM | gemm_universal [1][2] engine: gemm_universal/ example: 03_gemm/ |
✅ | ✅ | ✅ | ✅ | ❌ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ||
| GEMM | gemm_multi_d [3] engine: gemm_multi_d/ example: 19_gemm_multi_d/ |
✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ❌ | ||||||
| GEMM | gemm_preshuffle [4] engine: gemm_preshuffle/ |
✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ❌ | ||||||
| GEMM | streamk_gemm [5][6][7] engine: gemm_streamk/ example: 40_streamk_gemm/ |
✅ | ✅ | ❌ | ❌ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ | ❌ | ❌ | |||
| GEMM | batched_gemm example: 16_batched_gemm/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| GEMM | batched_contraction example: 41_batched_contraction/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| GEMM | block_scale_gemm example: 38_block_scale_gemm/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ||||||
| GEMM | flatmm example: 18_flatmm/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ||||
| GEMM | gemm_multi_abd example: 22_gemm_multi_abd/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| GEMM | gemm_quant | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ||||||||
| GEMM | grouped_gemm example: 17_grouped_gemm/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ||||
| GEMM | grouped_gemm_quant | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ||||||||
| Reduce | multi_reduce2d [8] engine: reduce/ example: 05_reduce/ |
✅ | ❌ | ❌ | ✅ | ✅ | ❌ | |||||||||
| Reduce | reduce2d example: 05_reduce/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Attention | fmha example: 01_fmha/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||
| Attention | sparse_attn example: 50_sparse_attn/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Activation | softmax | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Activation | topk_softmax example: 09_topk_softmax/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Conv | grouped_conv example: 20_grouped_convolution/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Data Move | batched_transpose example: 35_batched_transpose/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | ||||||||
| Data Move | image_to_column example: 04_img2col/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Data Move | permute example: 06_permute/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Elementwise | elementwise example: 21_elementwise/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| MoE | fused_moe example: 15_fused_moe/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Norm | add_rmsnorm2d_rdquant example: 11_add_rmsnorm2d_rdquant/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Norm | layernorm2d example: 02_layernorm2d/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Norm | norm_reduce | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Norm | rmsnorm2d example: 10_rmsnorm2d/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ | |||||||||
| Pooling | pooling example: 36_pooling/ |
❌ | ❌ | ❌ | ❌ | ❌ | ||||||||||
| Quant | smoothquant example: 12_smoothquant/ |
❌ | ❌ | ❌ | ❌ | ❌ | ❌ |
Legend:
- CK Tile Kernel column: First line is the kernel name. Lines prefixed with "engine:" show the tile engine directory under
ops/. Lines prefixed with "example:" show the CK Tile example directory underexample/ck_tile/. - Green cell (✅): CK Tile implementation exists and the tile engine supports it.
- Red cell (❌): CK Tile implementation exists but the tile engine does not support it.
- Grey cell (blank): No CK Tile implementation exists for this combination.
Notes:
- All CK Tile GEMM and reduce kernels are architecturally generic (no compile-time GPU guards). The gfx filtering in the tile engine is a validation/testing scope decision, not a code limitation.
- [1] gemm_universal: CMake defaults to
fp8;fp16. Building bf16/bf8 requires-DGEMM_UNIVERSAL_DATATYPE="fp16;fp8;bf16;bf8". - [2] gemm_universal: CK Tile supports int8 GEMM (with int32 output) but the tile engine has no int8 configuration.
- [3] gemm_multi_d: CK Tile kernel is type-generic but example and tile engine are fp16-only. Adding other types requires new tile engine configurations.
- [4] gemm_preshuffle: Only supports rcr layout (A=row, B=column, C=row) due to the pre-shuffle data format requirement.
- [5] streamk_gemm: CK Tile supports bf16 and bf8 for streamk, but the tile engine has no default tile configs for them.
- [6] streamk_gemm: Builder and default configs support all 4 layouts, but CMake defaults to
rcronly. Building others requires-DGEMM_STREAMK_LAYOUT="rcr;rrr;ccr;crr". - [7] streamk_gemm: CK Tile kernels have no arch-specific guards; tile engine filtering is pending validation for gfx950/gfx1201.
- [8] multi_reduce2d: CK Tile's reduce example supports bf16 input but the tile engine only configures fp16. The reduce kernel adapts to wave32/wave64 at runtime via
is_wave32(). - Reduce operations do not use matrix layouts.
Layout codes: Each layout code specifies the memory layout of tensors A, B, and C as row-major (r) or column-major (c). For example, rcr means A is row-major, B is column-major, and C is row-major. For gemm_multi_d, the instance builder uses 4-character codes (e.g., rcrr) where the 4th character specifies the D tensor layout; in the table above, the 3-character A/B/C portion is shown since the D layout is always row-major (r) for all supported configurations.
Data type mapping: The column labels (fp16, fp8, bf16, bf8, int8, fp4, fp6) refer to the input configuration label passed to the tile engine or CK Tile example. Each label determines the actual types used for the source tensors (A, B), accumulator, and output tensor (C). For 16-bit and 8-bit float types, A and B use the label type, the accumulator is fp32, and the output type C matches the input type for fp16 and bf16 but is promoted to fp16 for fp8 and bf8 since 8-bit precision is insufficient for output storage. int8 uses int32 for both accumulation and output. fp4 is a mixed-precision weight type where B is fp4 and A uses the activation type (fp16 or bf16). fp6 is used by the microscaling (MX) flatmm pipeline where both A and B are fp6 with fp32 accumulation and fp32 output.
Data type mapping per config label:
| Config Label | A (source) | B (source) | Acc | C (output) |
|---|---|---|---|---|
| fp16 | fp16 | fp16 | fp32 | fp16 |
| bf16 | bf16 | bf16 | fp32 | bf16 |
| int8 | int8 | int8 | int32 | int32 |
| fp8 | fp8 | fp8 | fp32 | fp16 |
| bf8 | bf8 | bf8 | fp32 | fp16 |
| fp6 | fp6 | fp6 | fp32 | fp32 |
| fp4 | fp16 or bf16 | fp4 | fp32 | fp16 or bf16 |
For gemm_multi_d, the D tensors (D0, D1) use the same type as the config label (fp16).