mirror of
https://github.com/NVIDIA/cutlass.git
synced 2026-05-11 17:00:05 +00:00
[CuTeDSL] Add dataclass example: passing pointers via frozen dataclass (#3070)
* Add dataclass example: passing pointers via frozen dataclass Demonstrates passing pointers from tensor arguments in @cute.jit to @cute.kernel using @dataclass(frozen=True). Shows the pattern of extracting pointers with tensor.iterator, bundling into a dataclass, and reconstructing tensors in the kernel. Uses fake tensors for compilation and TVM-FFI for runtime dispatch. Co-authored-by: Cursor <cursoragent@cursor.com> * Add dataclass example: passing tensors via frozen dataclass Demonstrates passing tensors from @cute.jit to @cute.kernel using @dataclass(frozen=True). Shows the pattern of bundling tensors into a dataclass with static configuration. Uses fake tensors for compilation and TVM-FFI for runtime dispatch. Includes reference check against PyTorch implementation. Co-authored-by: Cursor <cursoragent@cursor.com> --------- Co-authored-by: Cursor <cursoragent@cursor.com>
This commit is contained in:
98
examples/python/CuTeDSL/cute/dataclass_immutable.py
Normal file
98
examples/python/CuTeDSL/cute/dataclass_immutable.py
Normal file
@@ -0,0 +1,98 @@
|
||||
# Copyright (c) 2025 - 2026 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
||||
# SPDX-License-Identifier: BSD-3-Clause
|
||||
|
||||
"""
|
||||
Dataclass Example: Passing Tensors via Frozen Dataclass
|
||||
=======================================================
|
||||
|
||||
Demonstrates passing tensors from `@cute.jit` to `@cute.kernel`
|
||||
using `@dataclass(frozen=True)`.
|
||||
|
||||
Uses fake tensors for compilation and TVM-FFI for efficient runtime dispatch.
|
||||
|
||||
To run:
|
||||
|
||||
.. code-block:: bash
|
||||
|
||||
python examples/python/CuTeDSL/cute/dataclass_immutable.py
|
||||
|
||||
"""
|
||||
|
||||
from dataclasses import dataclass
|
||||
|
||||
import cutlass
|
||||
import cutlass.cute as cute
|
||||
from cutlass import Int32
|
||||
|
||||
|
||||
@dataclass(frozen=True)
|
||||
class CopyParams:
|
||||
"""
|
||||
Dataclass with one static attribute and two tensors.
|
||||
|
||||
Attributes:
|
||||
num_threads: Threads per block (static - used for loop unrolling)
|
||||
src: Source tensor (dynamic)
|
||||
dst: Destination tensor (dynamic)
|
||||
"""
|
||||
|
||||
num_threads: int # static
|
||||
src: cute.Tensor # dynamic
|
||||
dst: cute.Tensor # dynamic
|
||||
|
||||
|
||||
@cute.jit
|
||||
def memcpy_jit(mSrc: cute.Tensor, mDst: cute.Tensor, num_threads: cutlass.Constexpr[int]):
|
||||
"""Host function - passes tensors via dataclass."""
|
||||
params = CopyParams(
|
||||
num_threads=num_threads,
|
||||
src=mSrc,
|
||||
dst=mDst,
|
||||
)
|
||||
|
||||
N = mSrc.shape[0]
|
||||
num_blocks = cute.ceil_div(N, num_threads)
|
||||
|
||||
memcpy_kernel(params, N).launch(
|
||||
grid=[num_blocks, 1, 1],
|
||||
block=[num_threads, 1, 1],
|
||||
)
|
||||
|
||||
|
||||
@cute.kernel
|
||||
def memcpy_kernel(params: CopyParams, N: Int32):
|
||||
"""Device kernel - receives tensors via CopyParams dataclass."""
|
||||
tidx, _, _ = cute.arch.thread_idx()
|
||||
bidx, _, _ = cute.arch.block_idx()
|
||||
|
||||
idx = bidx * params.num_threads + tidx
|
||||
|
||||
if idx < N:
|
||||
params.dst[idx] = params.src[idx]
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
import torch
|
||||
from cutlass.cute.runtime import make_fake_compact_tensor
|
||||
|
||||
n = cute.sym_int64()
|
||||
# 1. Create fake tensors for compilation (no GPU memory allocated)
|
||||
fake_src = make_fake_compact_tensor(cutlass.Float32, (n,), stride_order=(0,))
|
||||
fake_dst = make_fake_compact_tensor(cutlass.Float32, (n,), stride_order=(0,))
|
||||
|
||||
# 2. Compile with --enable-tvm-ffi
|
||||
compiled_fn = cute.compile(
|
||||
memcpy_jit, fake_src, fake_dst, num_threads=128, options="--enable-tvm-ffi"
|
||||
)
|
||||
|
||||
# 3. Create real torch tensors
|
||||
src = torch.randn(1024, device="cuda", dtype=torch.float32)
|
||||
dst = torch.zeros_like(src)
|
||||
|
||||
# 4. Run compiled kernel
|
||||
compiled_fn(src, dst)
|
||||
|
||||
# 5. Verify dst == src
|
||||
torch.cuda.synchronize()
|
||||
torch.testing.assert_close(dst, src)
|
||||
print("PASS")
|
||||
Reference in New Issue
Block a user