mirror of
https://github.com/ROCm/composable_kernel.git
synced 2026-05-04 21:51:28 +00:00
Adding dispatcher architecture (#3300)
* WIP POC of dispatcher * Dispatcher python workflow setup. * Dispatcher cleanup and updates. Further dispatcher cleanup and updates. Build fixes Improvements and python to CK example Improvements to readme * Fixes to python paths * Cleaning up code * Improving dispatcher support for different arch Fixing typos * Fix formatting errors * Cleaning up examples * Improving codegeneration * Improving and fixing C++ examples * Adding conv functionality (fwd,bwd,bwdw) and examples. * Fixes based on feedback. * Further fixes based on feedback. * Adding stress test for autogeneration and autocorrection, and fixing preshuffle bug. * Another round of improvements based on feedback. * Trimming out unnecessary code. * Fixing the multi-D implementation. * Using gpu verification for gemms and fixing convolutions tflops calculation. * Fix counter usage issue and arch filtering per ops. * Adding changelog and other fixes. * Improve examples and resolve critical bugs. * Reduce build time for python examples. * Fixing minor bug. * Fix compilation error. * Improve installation instructions for dispatcher. * Add docker based installation instructions for dispatcher. * Fixing arch-based filtering to match tile engine. * Remove dead code and fix arch filtering. * Minor bugfix. * Updates after rebase. * Trimming code. * Fix copyright headers. * Consolidate examples, cut down code. * Minor fixes. * Improving python examples. * Update readmes. * Remove conv functionality. * Cleanup following conv removable.
This commit is contained in:
committed by
GitHub
parent
44f481a45c
commit
9e049a32a1
166
dispatcher/examples/gemm/python/05_numpy_integration.py
Normal file
166
dispatcher/examples/gemm/python/05_numpy_integration.py
Normal file
@@ -0,0 +1,166 @@
|
||||
#!/usr/bin/env python3
|
||||
|
||||
# Copyright (c) Advanced Micro Devices, Inc., or its affiliates.
|
||||
# SPDX-License-Identifier: MIT
|
||||
|
||||
"""
|
||||
Example 05: NumPy Integration
|
||||
|
||||
Shows how to create a GPU-accelerated matmul wrapper.
|
||||
|
||||
Complexity: ★★☆☆☆
|
||||
|
||||
Usage:
|
||||
python3 05_numpy_integration.py
|
||||
python3 05_numpy_integration.py --help
|
||||
python3 05_numpy_integration.py --dtype bf16
|
||||
"""
|
||||
|
||||
import sys
|
||||
import argparse
|
||||
from pathlib import Path
|
||||
|
||||
sys.path.insert(0, str(Path(__file__).parent.parent.parent.parent / "python"))
|
||||
import numpy as np
|
||||
|
||||
from ctypes_utils import (
|
||||
KernelConfig,
|
||||
Dispatcher,
|
||||
setup_gemm_dispatcher,
|
||||
cleanup_gemm,
|
||||
reset_for_example,
|
||||
)
|
||||
|
||||
|
||||
class GPUMatmul:
|
||||
"""GPU-accelerated matrix multiplication wrapper."""
|
||||
|
||||
def __init__(self, dispatcher: Dispatcher):
|
||||
self.dispatcher = dispatcher
|
||||
|
||||
def __call__(self, A: np.ndarray, B: np.ndarray) -> np.ndarray:
|
||||
"""Compute C = A @ B on GPU with CPU fallback."""
|
||||
M, K = A.shape
|
||||
K2, N = B.shape
|
||||
|
||||
if K != K2:
|
||||
raise ValueError(f"Dimension mismatch: {A.shape} @ {B.shape}")
|
||||
|
||||
if not self.dispatcher.is_supported(M, N, K):
|
||||
return np.matmul(A, B)
|
||||
|
||||
result = self.dispatcher.run(A, B, M, N, K)
|
||||
return result.output if result.success else np.matmul(A, B)
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(
|
||||
description="NumPy Integration Example - GPU-accelerated matmul wrapper",
|
||||
formatter_class=argparse.RawDescriptionHelpFormatter,
|
||||
epilog="""
|
||||
Examples:
|
||||
python3 05_numpy_integration.py # Default FP16
|
||||
python3 05_numpy_integration.py --dtype bf16 # BF16 mode
|
||||
""",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--dtype",
|
||||
default="fp16",
|
||||
choices=["fp16", "bf16", "fp32"],
|
||||
help="Data type (default: fp16)",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--arch", default="gfx942", help="Target architecture (default: gfx942)"
|
||||
)
|
||||
args = parser.parse_args()
|
||||
|
||||
reset_for_example()
|
||||
|
||||
print("=" * 60)
|
||||
print("Example 05: NumPy Integration")
|
||||
print("=" * 60)
|
||||
|
||||
# =========================================================================
|
||||
# Step 1: Setup dispatcher
|
||||
# =========================================================================
|
||||
print("\nStep 1: Setup Dispatcher")
|
||||
|
||||
config = KernelConfig(
|
||||
dtype_a=args.dtype,
|
||||
dtype_b=args.dtype,
|
||||
dtype_c=args.dtype,
|
||||
tile_m=128,
|
||||
tile_n=128,
|
||||
tile_k=32,
|
||||
gfx_arch=args.arch,
|
||||
)
|
||||
|
||||
setup = setup_gemm_dispatcher(config, registry_name="numpy", verbose=True)
|
||||
if not setup.success:
|
||||
print(f" ERROR: {setup.error}")
|
||||
return 1
|
||||
|
||||
dispatcher = setup.dispatcher
|
||||
np_dtype = np.float16 if args.dtype in ["fp16", "bf16"] else np.float32
|
||||
|
||||
# =========================================================================
|
||||
# Step 2: Create GPU matmul wrapper
|
||||
# =========================================================================
|
||||
print("\nStep 2: Create GPUMatmul")
|
||||
|
||||
gpu_matmul = GPUMatmul(dispatcher=dispatcher)
|
||||
print(" gpu_matmul ready")
|
||||
|
||||
# =========================================================================
|
||||
# Step 3: Demo - Simple multiplication using gpu_matmul
|
||||
# =========================================================================
|
||||
print("\nStep 3: Demo - Simple Multiplication")
|
||||
|
||||
A = np.random.randn(1024, 512).astype(np_dtype) * 0.1
|
||||
B = np.random.randn(512, 256).astype(np_dtype) * 0.1
|
||||
|
||||
# Use the gpu_matmul wrapper
|
||||
C = gpu_matmul(A, B)
|
||||
print(f" gpu_matmul result: {C.shape}, sum={C.sum():.4f}")
|
||||
|
||||
M, K = A.shape
|
||||
_, N = B.shape
|
||||
result = dispatcher.run(A, B, M, N, K)
|
||||
|
||||
print(f" A: {A.shape}, B: {B.shape} -> C: {result.output.shape}")
|
||||
print(f" GPU: {result.time_ms:.4f} ms, {result.tflops:.2f} TFLOPS")
|
||||
|
||||
# =========================================================================
|
||||
# Step 4: Demo - FFN block
|
||||
# =========================================================================
|
||||
print("\nStep 4: Demo - FFN Block")
|
||||
|
||||
batch, hidden, ffn = 128, 768, 3072
|
||||
X = np.random.randn(batch, hidden).astype(np_dtype) * 0.02
|
||||
W1 = np.random.randn(hidden, ffn).astype(np_dtype) * 0.02
|
||||
W2 = np.random.randn(ffn, hidden).astype(np_dtype) * 0.02
|
||||
|
||||
result1 = dispatcher.run(X, W1, batch, ffn, hidden)
|
||||
H = result1.output
|
||||
result2 = dispatcher.run(H, W2, batch, hidden, ffn)
|
||||
|
||||
print(f" X: {X.shape} -> H: {H.shape} -> Y: {result2.output.shape}")
|
||||
print(f" Total: {result1.time_ms + result2.time_ms:.4f} ms")
|
||||
|
||||
# Cleanup
|
||||
cleanup_gemm()
|
||||
|
||||
# Summary
|
||||
print("\n" + "=" * 60)
|
||||
print("NumPy Integration Pattern:")
|
||||
print("=" * 60)
|
||||
print(" 1. setup_gemm_dispatcher(config)")
|
||||
print(" 2. GPUMatmul(dispatcher)")
|
||||
print(" 3. C = gpu_matmul(A, B)")
|
||||
print("=" * 60)
|
||||
|
||||
return 0
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
sys.exit(main())
|
||||
Reference in New Issue
Block a user