mirror of
https://github.com/NVIDIA/nvbench.git
synced 2026-04-19 22:38:52 +00:00
Improvements to readability of examples per PR review
This commit is contained in:
@@ -21,11 +21,11 @@ import numpy as np
|
||||
from numba import cuda
|
||||
|
||||
|
||||
def as_cuda_Stream(cs: nvbench.CudaStream) -> cuda.cudadrv.driver.Stream:
|
||||
def as_cuda_stream(cs: nvbench.CudaStream) -> cuda.cudadrv.driver.Stream:
|
||||
return cuda.external_stream(cs.addressof())
|
||||
|
||||
|
||||
def make_kernel(items_per_thread: int) -> cuda.dispatcher.CUDADispatcher:
|
||||
def make_throughput_kernel(items_per_thread: int) -> cuda.dispatcher.CUDADispatcher:
|
||||
@cuda.jit
|
||||
def kernel(stride: np.uintp, elements: np.uintp, in_arr, out_arr):
|
||||
tid = cuda.grid(1)
|
||||
@@ -46,7 +46,7 @@ def throughput_bench(state: nvbench.State) -> None:
|
||||
nbytes = 128 * 1024 * 1024
|
||||
elements = nbytes // np.dtype(np.int32).itemsize
|
||||
|
||||
alloc_stream = as_cuda_Stream(state.get_stream())
|
||||
alloc_stream = as_cuda_stream(state.get_stream())
|
||||
inp_arr = cuda.device_array(elements, dtype=np.int32, stream=alloc_stream)
|
||||
out_arr = cuda.device_array(elements * ipt, dtype=np.int32, stream=alloc_stream)
|
||||
|
||||
@@ -56,7 +56,7 @@ def throughput_bench(state: nvbench.State) -> None:
|
||||
threads_per_block = 256
|
||||
blocks_in_grid = (elements + threads_per_block - 1) // threads_per_block
|
||||
|
||||
krn = make_kernel(ipt)
|
||||
krn = make_throughput_kernel(ipt)
|
||||
|
||||
# warm-up call ensures that kernel is loaded into context
|
||||
# before blocking kernel is launched. Kernel loading may cause
|
||||
@@ -66,7 +66,7 @@ def throughput_bench(state: nvbench.State) -> None:
|
||||
)
|
||||
|
||||
def launcher(launch: nvbench.Launch):
|
||||
exec_stream = as_cuda_Stream(launch.get_stream())
|
||||
exec_stream = as_cuda_stream(launch.get_stream())
|
||||
krn[blocks_in_grid, threads_per_block, exec_stream, 0](
|
||||
stride, elements, inp_arr, out_arr
|
||||
)
|
||||
|
||||
@@ -193,7 +193,7 @@ if __name__ == "__main__":
|
||||
nvbench.register(default_value)
|
||||
# specify axis
|
||||
nvbench.register(single_float64_axis).add_float64_axis(
|
||||
"Duration", [7e-5, 1e-4, 5e-4]
|
||||
"Duration (s)", [7e-5, 1e-4, 5e-4]
|
||||
)
|
||||
|
||||
copy1_bench = nvbench.register(copy_sweep_grid_shape)
|
||||
|
||||
@@ -61,10 +61,6 @@ def segmented_reduce(state: nvbench.State):
|
||||
dev_id = state.get_device()
|
||||
cp_stream = as_cp_ExternalStream(state.get_stream(), dev_id)
|
||||
|
||||
with cp_stream:
|
||||
rng = cp.random.default_rng()
|
||||
mat = rng.integers(low=-31, high=32, dtype=np.int32, size=(n_rows, n_cols))
|
||||
|
||||
def add_op(a, b):
|
||||
return a + b
|
||||
|
||||
@@ -84,6 +80,8 @@ def segmented_reduce(state: nvbench.State):
|
||||
|
||||
h_init = np.zeros(tuple(), dtype=np.int32)
|
||||
with cp_stream:
|
||||
rng = cp.random.default_rng()
|
||||
mat = rng.integers(low=-31, high=32, dtype=np.int32, size=(n_rows, n_cols))
|
||||
d_input = mat
|
||||
d_output = cp.empty(n_rows, dtype=d_input.dtype)
|
||||
|
||||
|
||||
@@ -36,14 +36,16 @@ def cupy_extract_by_mask(state: nvbench.State):
|
||||
|
||||
state.collect_cupti_metrics()
|
||||
state.add_element_count(n_rows * n_cols, "# Elements")
|
||||
int32_dt = cp.dtype(cp.int32)
|
||||
bool_dt = cp.dtype(cp.bool_)
|
||||
state.add_global_memory_reads(
|
||||
n_rows * n_cols * (cp.dtype(cp.int32).itemsize + cp.dtype("?").itemsize)
|
||||
n_rows * n_cols * (int32_dt.itemsize + bool_dt.itemsize)
|
||||
)
|
||||
state.add_global_memory_writes(n_rows * n_cols * (cp.dtype(cp.int32).itemsize))
|
||||
state.add_global_memory_writes(n_rows * n_cols * (int32_dt.itemsize))
|
||||
|
||||
with cp_s:
|
||||
X = cp.full((n_cols, n_rows), fill_value=3, dtype=cp.int32)
|
||||
mask = cp.ones((n_cols, n_rows), dtype="?")
|
||||
X = cp.full((n_cols, n_rows), fill_value=3, dtype=int32_dt)
|
||||
mask = cp.ones((n_cols, n_rows), dtype=bool_dt)
|
||||
_ = X[mask]
|
||||
|
||||
def launcher(launch: nvbench.Launch):
|
||||
|
||||
@@ -21,11 +21,11 @@ import numpy as np
|
||||
from numba import cuda
|
||||
|
||||
|
||||
def as_cuda_Stream(cs: nvbench.CudaStream) -> cuda.cudadrv.driver.Stream:
|
||||
def as_cuda_stream(cs: nvbench.CudaStream) -> cuda.cudadrv.driver.Stream:
|
||||
return cuda.external_stream(cs.addressof())
|
||||
|
||||
|
||||
def make_kernel(items_per_thread: int) -> cuda.dispatcher.CUDADispatcher:
|
||||
def make_throughput_kernel(items_per_thread: int) -> cuda.dispatcher.CUDADispatcher:
|
||||
@cuda.jit
|
||||
def kernel(stride: np.uintp, elements: np.uintp, in_arr, out_arr):
|
||||
tid = cuda.grid(1)
|
||||
@@ -46,7 +46,7 @@ def throughput_bench(state: nvbench.State) -> None:
|
||||
nbytes = 128 * 1024 * 1024
|
||||
elements = nbytes // np.dtype(np.int32).itemsize
|
||||
|
||||
alloc_stream = as_cuda_Stream(state.get_stream())
|
||||
alloc_stream = as_cuda_stream(state.get_stream())
|
||||
inp_arr = cuda.device_array(elements, dtype=np.int32, stream=alloc_stream)
|
||||
out_arr = cuda.device_array(elements * ipt, dtype=np.int32, stream=alloc_stream)
|
||||
|
||||
@@ -57,7 +57,7 @@ def throughput_bench(state: nvbench.State) -> None:
|
||||
threads_per_block = 256
|
||||
blocks_in_grid = (elements + threads_per_block - 1) // threads_per_block
|
||||
|
||||
krn = make_kernel(ipt)
|
||||
krn = make_throughput_kernel(ipt)
|
||||
|
||||
# warm-up call ensures that kernel is loaded into context
|
||||
# before blocking kernel is launched. Kernel loading may
|
||||
@@ -67,7 +67,7 @@ def throughput_bench(state: nvbench.State) -> None:
|
||||
)
|
||||
|
||||
def launcher(launch: nvbench.Launch):
|
||||
exec_stream = as_cuda_Stream(launch.get_stream())
|
||||
exec_stream = as_cuda_stream(launch.get_stream())
|
||||
krn[blocks_in_grid, threads_per_block, exec_stream, 0](
|
||||
stride, elements, inp_arr, out_arr
|
||||
)
|
||||
|
||||
Reference in New Issue
Block a user