mirror of
https://github.com/ROCm/composable_kernel.git
synced 2026-03-19 22:57:38 +00:00
2D Pooling Forward
Theory
This example demonstrates the 2D pooling forward pass, a key operation in convolutional neural networks (CNNs) for spatial downsampling. Pooling reduces the spatial dimensions of feature maps, providing translation invariance and reducing computation.
Mathematical Formulation:
Given input X[N, C, H_{in}, W_{in}], pooling window (k_H, k_W), stride (s_H, s_W), and padding (p_H, p_W):
- Output
Y[N, C, H_{out}, W_{out}] H_{out} = \left\lfloor \frac{H_{in} + 2p_H - k_H}{s_H} \right\rfloor + 1W_{out} = \left\lfloor \frac{W_{in} + 2p_W - k_W}{s_W} \right\rfloor + 1
For each output position:
- Max Pooling:
Y_{n,c,h,w} = \max_{i,j} X_{n,c,h \cdot s_H + i, w \cdot s_W + j} - Average Pooling:
Y_{n,c,h,w} = \frac{1}{k_H k_W} \sum_{i,j} X_{n,c,h \cdot s_H + i, w \cdot s_W + j}
Algorithmic Background:
- Each thread computes one or more output elements.
- Handles padding and boundary conditions.
- Optimizes memory access for bandwidth.
How to Run
Prerequisites
Please follow the instructions in the main Build Guide section as a prerequisite to building and running this example.
Build and run
cd composable_kernel/example/13_pool2d_fwd
mkdir build && cd build
cmake -DCMAKE_CXX_COMPILER=/opt/rocm/bin/hipcc ..
make -j
Run example_pool2d_fwd_fp16
#arg1: verification (0=no, 1=yes)
#arg2: initialization (0=no init, 1=single integer value, 2=scope integer value, 3=decimal value)
#arg3: time kernel (0=no, 1=yes)
#arg4 to 15: N, C, Y, X, Hi, Wi, Sy, Sx, LeftPy, LeftPx, RightPy, RightPx
./bin/example_pool2d_fwd_fp16 1 1 1
Expected Result:
in_n_c_hi_wi: dim 4, lengths {128, 192, 71, 71}, strides {967872, 1, 13632, 192}
out_n_c_ho_wo: dim 4, lengths {128, 192, 36, 36}, strides {248832, 1, 6912, 192}
launch_and_time_kernel: grid_dim {124416, 1, 1}, block_dim {64, 1, 1}
Warm up 1 time
Start running 10 times...
Perf: 0.397436 ms, 1.44252 TFlops, 783.713 GB/s
Run example_pool2d_fwd_fp32
#arg1: verification (0=no, 1=yes)
#arg2: initialization (0=no init, 1=single integer value, 2=scope integer value, 3=decimal value)
#arg3: time kernel (0=no, 1=yes)
#arg4 to 15: N, C, Y, X, Hi, Wi, Sy, Sx, LeftPy, LeftPx, RightPy, RightPx
./bin/example_pool2d_fwd_fp32 1 1 1
Expected Result:
./bin/example_pool2d_fwd_fp32 1 1 1
in_n_c_hi_wi: dim 4, lengths {128, 192, 71, 71}, strides {967872, 1, 13632, 192}
out_n_c_ho_wo: dim 4, lengths {128, 192, 36, 36}, strides {248832, 1, 6912, 192}
launch_and_time_kernel: grid_dim {124416, 1, 1}, block_dim {64, 1, 1}
Warm up 1 time
Start running 10 times...
Perf: 1.01823 ms, 0.563045 TFlops, 611.8 GB/s
Source Code Structure
Directory Layout
example/13_pool2d_fwd/
├── pool2d_fwd_xdl.cpp # Main example: sets up, runs, and verifies 2D pooling
include/ck/tensor_operation/gpu/device/
│ └── device_pool_fwd.hpp # Device-level pooling API
include/ck/tensor_operation/gpu/device/impl/
│ └── device_pool2d_fwd_nhwc.hpp # NHWC layout optimization
│ └── device_pool2d_fwd_nchw.hpp # NCHW layout optimization
include/ck/tensor_operation/gpu/grid/
│ └── gridwise_pool_fwd.hpp # Grid-level pooling kernel
include/ck/tensor_operation/gpu/block/
└── blockwise_pool.hpp # Block-level pooling
Key Classes and Functions
- DevicePoolFwd (in
device_pool_fwd.hpp):
Device API for pooling. - gridwise_pool_fwd (in
gridwise_pool_fwd.hpp):
Implements the tiled/blocking pooling kernel. - blockwise_pool (in
blockwise_pool.hpp):
Handles block-level pooling and shared memory.
This example demonstrates how Composable Kernel implements efficient 2D pooling for CNNs and vision models.