Files

Client Example: GEMM with Add, Add, and FastGELU Fusion

Theory

This client example demonstrates GEMM fused with two addition operations and FastGELU activation. This pattern is common in transformer feed-forward networks and other neural architectures where a linear transformation is followed by bias addition, residual addition, and a non-linear activation.

Mathematical Formulation:


E = \text{FastGELU}((A \times B) + D_0 + D_1)
  • A: [M, K] input matrix
  • B: [K, N] weight matrix
  • D_0: [N] bias vector (broadcasted)
  • D_1: [M, N] residual tensor
  • E: [M, N] output

FastGELU is an efficient approximation of GELU:


\text{FastGELU}(x) = x \cdot \sigma(1.702 \cdot x)

where \sigma is the sigmoid function.

Algorithmic Background:

  • The GEMM result is kept in registers, bias and residual are added, and FastGELU is applied before writing to global memory.
  • No intermediate results are written to global memory.

How to Run

Prerequisites

Please follow the instructions in the main Build Guide section as a prerequisite to building and running this example.

Build and run

cd composable_kernel/client_example/02_gemm_add_add_fastgelu
mkdir build && cd build
cmake -DCMAKE_CXX_COMPILER=/opt/rocm/bin/hipcc ..
make -j

# Example run
./gemm_add_add_fastgelu

Source Code Structure

Directory Layout

client_example/02_gemm_add_add_fastgelu/
├── gemm_add_add_fastgelu.cpp         # Main client example: GEMM+Add+Add+FastGELU
├── gemm_add_add_fastgelu_generic.cpp # Generic variant
├── gemm_add_fastgelu.cpp             # GEMM+Add+FastGELU
├── gemm_add_fastgelu_generic.cpp     # Generic variant
├── gemm_fastgelu.cpp                 # GEMM+FastGELU only
├── gemm_fastgelu_generic.cpp         # Generic variant
├── CMakeLists.txt                    # Build configuration for the example

Key Functions

  • main() (in each .cpp):
    Sets up input matrices, configures GEMM and epilogue parameters, launches the fused kernel, and verifies the result.
  • Fused kernel invocation:
    Uses the Composable Kernel device API to launch the GEMM with fused addition and FastGELU.

This client example provides several variants to demonstrate different levels of fusion and genericity for transformer-style MLP layers.