# CK Tile Dispatcher A unified kernel dispatch system for AMD GPUs with C++ and Python frontends. **Validated Platform:** AMD Instinct MI300 series (gfx942) --- ## Table of Contents 1. [Quick Start](#quick-start) 2. [Docker Setup](#docker-setup-recommended) 3. [Prerequisites](#prerequisites) 4. [Step-by-Step Build Guide](#step-by-step-build-guide) 5. [Running Examples](#running-examples) 6. [External Integration](#external-integration) 7. [Core Concepts](#core-concepts) 8. [Troubleshooting](#troubleshooting) 9. [File Structure](#file-structure) --- ## Quick Start **Complete setup from scratch (5 minutes):** ```bash # From the composable_kernel root directory cd dispatcher # Step 1: Create build directory mkdir -p build && cd build # Step 2: Configure CMake cmake .. \ -DCMAKE_PREFIX_PATH=/opt/rocm \ -DCMAKE_CXX_COMPILER=/opt/rocm/bin/hipcc \ -DCMAKE_BUILD_TYPE=Release \ -DGPU_TARGETS="gfx942" \ -DBUILD_DISPATCHER_EXAMPLES=ON # Step 3: Generate kernels and build (CMake handles this automatically) make -j$(nproc) # Step 4: Run C++ examples ./examples/gemm_01_basic # Step 5: Build Python libraries (required for Python examples) make python_libs # Step 6: Run Python examples (from dispatcher directory) cd .. python3 examples/gemm/python/01_basic_gemm.py ``` --- ## Docker Setup (Recommended) For a reproducible build environment, use the official ROCm Docker image: ### Step 1: Pull and Run Container ```bash # Pull the CK Docker image docker pull rocm/composable_kernel:ck_ub24.04_rocm7.0.1 # Run container with GPU access docker run \ -it \ --privileged \ --device=/dev/kfd \ --device=/dev/dri \ --group-add video \ --group-add render \ -w /root/workspace \ -v $(pwd):/root/workspace \ rocm/composable_kernel:ck_ub24.04_rocm7.0.1 \ /bin/bash ``` > **Note:** Omit `--device` flags if building without GPU access. ### Step 2: Clone and Build ```bash # Inside the container git clone https://github.com/ROCm/composable_kernel.git cd composable_kernel git checkout builder-dispatch-tile-gemm # Set up Python environment python3 -m venv .venv source .venv/bin/activate pip install numpy # Build dispatcher cd dispatcher mkdir -p build && cd build cmake .. \ -DCMAKE_PREFIX_PATH=/opt/rocm \ -DCMAKE_CXX_COMPILER=/opt/rocm/bin/hipcc \ -DCMAKE_BUILD_TYPE=Release \ -DGPU_TARGETS="gfx942" \ -DBUILD_DISPATCHER_EXAMPLES=ON make -j$(nproc) ``` ### One-Liner Build (inside container) ```bash git clone https://github.com/ROCm/composable_kernel.git && \ cd composable_kernel && git checkout builder-dispatch-tile-gemm && \ python3 -m venv .venv && source .venv/bin/activate && pip install numpy && \ cd dispatcher && mkdir -p build && cd build && \ cmake .. -DCMAKE_PREFIX_PATH=/opt/rocm -DCMAKE_CXX_COMPILER=/opt/rocm/bin/hipcc \ -DCMAKE_BUILD_TYPE=Release -DGPU_TARGETS="gfx942" -DBUILD_DISPATCHER_EXAMPLES=ON && \ make -j$(nproc) ``` --- ## Prerequisites ### Required Software | Software | Minimum Version | Check Command | |----------|-----------------|---------------| | ROCm | 6.4+ | `rocminfo` | | CMake | 3.16+ | `cmake --version` | | Python | 3.8+ | `python3 --version` | | NumPy | 1.20+ | `pip show numpy` | | hipcc | (from ROCm) | `/opt/rocm/bin/hipcc --version` | > **Note:** Newer GPU targets (gfx950, gfx1201) require ROCm 6.3+. For ROCm 6.4+, you can also use `amdclang++` instead of `hipcc`. ### Check Your GPU Architecture ```bash # Find your GPU architecture rocminfo | grep -i "gfx" # Example output: "gfx942" ``` **Supported architectures:** - **gfx942** - MI300X, MI300A, MI308, MI325 (Instinct MI300 series) - **gfx90a** - MI200 series (MI250, MI250X) - **gfx950** - MI350 series - **gfx1101** - RDNA3 series - **gfx1201** - RDNA4 series ### Install Python Dependencies NumPy is required for Python examples and kernel generation. We recommend using a virtual environment: **Option 1: Using standard venv** ```bash # Create virtual environment python3 -m venv .venv # Activate virtual environment source .venv/bin/activate # Linux/macOS # .venv\Scripts\activate # Windows # Install NumPy pip install numpy ``` **Option 2: Using uv (faster alternative)** ```bash # Install uv if not already installed curl -LsSf https://astral.sh/uv/install.sh | sh # Create and activate virtual environment uv venv .venv source .venv/bin/activate # Linux/macOS # .venv\Scripts\activate # Windows # Install NumPy uv pip install numpy ``` **Option 3: System-wide install (not recommended)** ```bash pip install numpy ``` > **Note:** Always activate your virtual environment before running CMake or Python examples. ### Supported Data Types CK Tile supports a wide range of data types for GEMM operations: | A dtype | B dtype | Acc dtype | Warp Tile Sizes | Notes | |---------|---------|-----------|-----------------|-------| | `fp32` | `fp32` | `fp32` | 16x16x4, 16x16x16 | Full precision | | `fp16` | `fp16` | `fp32` | 32x32x8, 32x32x16, 16x16x16, 16x16x32 | Standard half | | `bf16` | `bf16` | `fp32` | 32x32x8, 32x32x16, 16x16x16, 16x16x32 | Brain float 16 | | `fp8` | `fp8` | `fp32` | 32x32x16, 32x32x32, 16x16x32, 16x16x64 | FP8 E4M3 | | `fp8` | `bf8` | `fp32` | 32x32x16, 16x16x32 | Mixed FP8/BF8 | | `bf8` | `fp8` | `fp32` | 32x32x16, 16x16x128 | Mixed BF8/FP8 | | `bf8` | `bf8` | `fp32` | 32x32x16, 32x32x32, 16x16x32 | BF8 E5M2 | | `int8` | `int8` | `int32` | 32x32x16, 16x16x32, 16x16x16 | Integer GEMM | | `pk_fp4` | `pk_fp4` | `fp32` | 16x16x128 | Packed 4-bit float | **Notes:** - Accumulator is always `fp32` except for `int8` which uses `int32` - FP8 types: `fp8` = E4M3, `bf8` = E5M2 - `pk_fp4` = Packed 4-bit float (2 values per byte) - Some dtypes require specific GPU architectures (e.g., FP8 requires MI300+) --- ## Step-by-Step Build Guide ### Step 1: Navigate to Dispatcher Directory ```bash # From composable_kernel root cd dispatcher # Verify you're in the right place ls CMakeLists.txt # Should exist ``` ### Step 2: Create Build Directory ```bash mkdir -p build cd build ``` ### Step 3: Configure CMake **Basic configuration (library only):** ```bash cmake .. \ -DCMAKE_PREFIX_PATH=/opt/rocm \ -DCMAKE_CXX_COMPILER=/opt/rocm/bin/hipcc \ -DCMAKE_BUILD_TYPE=Release \ -DGPU_TARGETS="gfx942" ``` **Full configuration (with examples and tests):** ```bash cmake .. \ -DCMAKE_PREFIX_PATH=/opt/rocm \ -DCMAKE_CXX_COMPILER=/opt/rocm/bin/hipcc \ -DCMAKE_BUILD_TYPE=Release \ -DGPU_TARGETS="gfx942" \ -DBUILD_DISPATCHER_EXAMPLES=ON \ -DBUILD_DISPATCHER_TESTS=ON ``` **Expected output:** ``` -- Found hip: /opt/rocm (found suitable version "6.x.x") -- Generating GEMM kernels... -- Built: gemm_01 through gemm_06, dispatcher_gemm_lib.so -- Configuring done ``` ### Step 4: Build ```bash # Build all targets (generates kernels automatically, then compiles) make -j$(nproc) # Or build specific targets make gemm_01_basic # Single GEMM example make dispatcher_gemm_lib # GEMM shared library for Python # Build ONLY Python libraries (faster if you don't need C++ examples) make python_libs -j$(nproc) ``` ### Kernel Generation Targets Kernels are generated automatically during `make`, but you can also control generation explicitly: ```bash # Generate all kernels only (no compilation) make generate_all_kernels # Generate GEMM kernels only make generate_gemm_kernels # Force regenerate (even if kernels exist) make regenerate_all_kernels make regenerate_gemm_kernels # Generate for specific GPU architecture make generate_kernels_gfx942 # MI300X make generate_kernels_gfx90a # MI200 make generate_kernels_gfx1100 # RDNA3 ``` ### Step 5: Verify Build ```bash # Check executables were built ls examples/gemm_* # Check shared libraries were built ls examples/libdispatcher_gemm_lib.so ``` ### CMake Options Reference | Flag | Default | Description | |------|---------|-------------| | `CMAKE_BUILD_TYPE` | Debug | **Use `Release` for performance!** | | `GPU_TARGETS` | None | Target GPU: `"gfx942"`, `"gfx90a"`, etc. | | `BUILD_DISPATCHER_EXAMPLES` | OFF | Build C++ examples and Python libs | | `BUILD_DISPATCHER_TESTS` | OFF | Build unit tests | | `CMAKE_PREFIX_PATH` | - | ROCm installation path | | `CMAKE_CXX_COMPILER` | - | Path to hipcc compiler | ⚠️ **Important:** Always use `-DCMAKE_BUILD_TYPE=Release` for benchmarking. Debug builds are slower. ⚠️ **Important:** Note that the current system provides single GPU target support for architecture-based kernel filtering, please do not use multiple GPU targets at a time (if necessary, please compile into different build directories). --- ## Running Examples ### C++ Examples After building, executables are in `build/examples/`: ```bash cd build/examples # GEMM Examples ./gemm_01_basic # Basic GEMM with autofill/autocorrect ./gemm_02_multi_size # Wildcard expansion ./gemm_03_benchmark_validation # Benchmarking + validation ./gemm_04_heuristics # Heuristic kernel selection ./gemm_05_json_export # Registry JSON export ./gemm_06_multi_registry # Multiple registries ``` ### Python Examples Run from the `dispatcher` directory: ```bash cd /path/to/composable_kernel/dispatcher # GEMM Examples python3 examples/gemm/python/01_basic_gemm.py # Basic multi-kernel GEMM python3 examples/gemm/python/04_validation.py # CPU reference validation python3 examples/gemm/python/07_stress_test.py # Stress test (48 kernels) python3 examples/gemm/python/08_heuristics.py # Heuristic selection ``` ### Example Output **Expected C++ output (`gemm_01_basic`):** ``` ====================================================================== Example 01: Basic GEMM with Declarative Kernel Definition ====================================================================== Step 1: Declared Kernels ------------------------ Kernel Set: fp16_gemm_kernels Architecture: gfx942 Configurations: 1 - gemm_fp16_rcr_compv4_cshuffle_intrawave_128x128x32 Step 2: Create Registry and Dispatcher -------------------------------------- Registered 1 kernels Step 3: Define Problem ---------------------- M=1024, N=1024, K=1024 Step 4: GPU Execution --------------------- *** GPU EXECUTION *** Time: ms TFLOPS: ``` > **Note:** Timing values vary by GPU model and system configuration. --- ## Benchmark Parameters The dispatcher supports fine-grained control over benchmarking, matching CK Tile's `stream_config`: ### Available Parameters | Parameter | Type | Default | Description | |-----------|------|---------|-------------| | `warmup` | int | 5 | Warmup iterations (discarded from timing) | | `repeat` | int | 20 | Benchmark iterations (averaged) | | `flush_cache` | bool | false | Flush GPU L2 cache between iterations | | `rotating_count` | int | 1 | Rotating buffer count (for cache simulation) | | `timer` | string | "gpu" | Timer type: "gpu" (HIP events) or "cpu" | | `init` | string | "random" | Matrix initialization: "random", "linear", "constant" | | `split_k` | int | 1 | Split-K parallelism factor | ### Python Usage ```python from ctypes_utils import DispatcherLib # Basic usage (default benchmark settings) lib = DispatcherLib.load() # Advanced benchmark settings via command line python3 examples/gemm/python/10_advanced_benchmark.py \ --warmup 10 \ --repeat 100 \ --flush-cache ``` ### C++ Usage ```cpp // Basic timing ck_tile::stream_config cfg{nullptr, true}; // Advanced benchmark settings ck_tile::stream_config cfg{ nullptr, // stream_id (nullptr = default stream) true, // time_kernel 1, // log_level 10, // cold_niters (warmup) 100, // nrepeat true, // is_gpu_timer true, // flush_cache 4 // rotating_count }; float avg_time = kernel.run(args, cfg); ``` ### Command Line (Python Examples) ```bash # Basic run python3 examples/gemm/python/10_advanced_benchmark.py # With benchmark parameters python3 examples/gemm/python/10_advanced_benchmark.py \ --warmup 10 \ --repeat 100 \ --flush-cache \ --rotating-count 4 \ --timer gpu ``` ### When to Use Each Parameter | Use Case | Recommended Settings | |----------|---------------------| | Quick test | `warmup=1, repeat=3` | | Stable benchmark | `warmup=10, repeat=100` | | Memory-bound analysis | `flush_cache=True, rotating_count=4` | | Compute-bound analysis | `flush_cache=False` (default) | | Debug timing | `timer="cpu"` | | Production | `timer="gpu"` (default) | --- ## External Integration ### Using Dispatcher in Your Own Project #### Option 1: CMake Integration (Recommended) Add to your `CMakeLists.txt`: ```cmake # Set path to composable_kernel set(CK_ROOT "/path/to/composable_kernel") # Add dispatcher subdirectory add_subdirectory(${CK_ROOT}/dispatcher dispatcher_build) # Link to your target target_link_libraries(your_target PRIVATE ck_tile_dispatcher) target_include_directories(your_target PRIVATE ${CK_ROOT}/dispatcher/include ${CK_ROOT}/include ) ``` #### Option 2: Include as Pre-built Library ```cmake # Find the pre-built library find_library(CK_DISPATCHER ck_tile_dispatcher PATHS /path/to/composable_kernel/dispatcher/build) # Include directories set(CK_INCLUDE_DIRS /path/to/composable_kernel/include /path/to/composable_kernel/dispatcher/include ) target_link_libraries(your_target PRIVATE ${CK_DISPATCHER}) target_include_directories(your_target PRIVATE ${CK_INCLUDE_DIRS}) ``` #### Option 3: Python Integration ```python import sys sys.path.insert(0, "/path/to/composable_kernel/dispatcher/examples/gemm/python") # For GEMM from ctypes_utils import DispatcherLib, Dispatcher, KernelConfig ``` ### Required Include Paths When integrating, you need these include paths: ``` /path/to/composable_kernel/include # CK Tile core headers /path/to/composable_kernel/dispatcher/include # Dispatcher headers /path/to/composable_kernel/dispatcher/build/generated_kernels # Generated kernels ``` ### Required Compile Flags ```bash # Minimum flags for hipcc -std=c++17 -D__HIP_PLATFORM_AMD__=1 --offload-arch=gfx942 # Your target GPU # Recommended flags -O3 -mllvm -enable-noalias-to-md-conversion=0 -Wno-undefined-func-template -Wno-float-equal -Wall -Werror ``` ### Python Path Setup For Python scripts outside the dispatcher directory: ```bash # Option 1: Environment variable export PYTHONPATH="/path/to/composable_kernel/dispatcher/examples/gemm/python:$PYTHONPATH" # Option 2: In your Python script import sys sys.path.insert(0, "/path/to/composable_kernel/dispatcher/examples/gemm/python") ``` ### Library Search Paths The Python utilities search for the shared library in these locations: ```python # For GEMM (ctypes_utils.py) SEARCH_PATHS = [ "build/examples/libdispatcher_gemm_lib.so", "../build/examples/libdispatcher_gemm_lib.so", "../../build/examples/libdispatcher_gemm_lib.so", ] ``` If using from a different location, set the library path explicitly: ```python # GEMM from ctypes_utils import DispatcherLib lib = DispatcherLib.load("/absolute/path/to/libdispatcher_gemm_lib.so") ``` --- ## Core Concepts ### Data Flow ``` KernelConfig → Registry → Dispatcher → GPU Execution ``` 1. **KernelConfig**: Defines kernel parameters (tile sizes, data types, layouts) 2. **Registry**: Stores multiple kernel configurations 3. **Dispatcher**: Selects best kernel for a given problem and executes it ### GEMM Layouts | Layout | A | B | C | Use Case | |--------|---|---|---|----------| | RCR | Row | Col | Row | Most common (PyTorch default) | | RRR | Row | Row | Row | Both inputs row-major | | CRR | Col | Row | Row | A transposed | | CCR | Col | Col | Row | Both inputs column-major | ### Split-K Support Split-K divides the K dimension across multiple thread blocks, useful for large K dimensions. **Usage (C++):** ```cpp // GEMM with 4-way K split auto problem = ProblemBuilder() .m(1024).n(1024).k(8192) .split_k(4) .build(); ``` --- ## Troubleshooting ### Build Issues | Problem | Solution | |---------|----------| | `hipcc not found` | Set `-DCMAKE_CXX_COMPILER=/opt/rocm/bin/hipcc` | | `hip not found` | Set `-DCMAKE_PREFIX_PATH=/opt/rocm` | | Very slow performance | Use `-DCMAKE_BUILD_TYPE=Release` | | `gfx942 not supported` | Check ROCm version (need 6.0+) | | Kernel generation fails | Ensure Python 3.8+ with NumPy installed in active venv | | Build errors | First verify CK builds without dispatcher (see main CK README) | ### Runtime Issues | Problem | Solution | |---------|----------| | `Library not found` | Build with `-DBUILD_DISPATCHER_EXAMPLES=ON` | | `No kernel found` | Check GPU arch matches build target | | Python `ModuleNotFoundError` | Add paths to `PYTHONPATH` (see above) | | Wrong results | Verify layout matches your data | ### Debug Commands ```bash # Check ROCm installation rocminfo | head -20 # Check GPU architecture rocminfo | grep "Name:" # Verify library exists ls -la build/examples/libdispatcher_*.so # Run with verbose output ./build/examples/gemm_01_basic 2>&1 # Python: Check library loading python3 -c " import ctypes lib = ctypes.CDLL('/path/to/libdispatcher_gemm_lib.so') print('Library loaded successfully') " ``` ### Clean Rebuild If you encounter issues, try a clean rebuild: ```bash cd dispatcher rm -rf build mkdir build && cd build cmake .. [your options] make -j$(nproc) ``` --- ## File Structure ``` dispatcher/ ├── README.md # This file ├── CMakeLists.txt # Build configuration │ ├── include/ck_tile/dispatcher/ # C++ headers │ ├── dispatcher.hpp # GEMM dispatcher │ ├── registry.hpp # Kernel registry │ └── kernel_key.hpp # Kernel configuration │ ├── src/ # C++ implementation │ ├── codegen/ # Kernel generation │ ├── unified_gemm_codegen.py # GEMM kernel generator │ └── arch_specs.json # GPU specifications │ ├── bindings/ctypes/ # Python ctypes interface │ └── gemm_ctypes_lib.cpp # GEMM Python library │ ├── examples/ # Examples │ └── gemm/ │ ├── cpp/ # C++ GEMM examples (01-06) │ └── python/ # Python GEMM examples (01-11) │ ├── scripts/ # Build scripts │ └── tests/ # Unit tests ``` --- ## Example Documentation | Directory | README | |-----------|--------| | GEMM C++ | [examples/gemm/cpp/README.md](examples/gemm/cpp/README.md) | | GEMM Python | [examples/gemm/python/README.md](examples/gemm/python/README.md) | | Codegen | [codegen/README.md](codegen/README.md) | --- ## Archived Content Convolution examples and utilities have been archived to `ck-2/conv_archive/dispatcher/`: - `examples/conv/cpp/` - 11 C++ convolution examples - `examples/conv/python/` - 14 Python convolution examples - `codegen/unified_conv_codegen.py` - Conv kernel generator - `include/ck_tile/dispatcher/conv_*.hpp` - Conv headers - `python/conv_utils.py` - Conv Python utilities --- ## License MIT License - Copyright (c) 2025, Advanced Micro Devices, Inc.