mirror of
https://github.com/NVIDIA/nvbench.git
synced 2026-03-14 20:27:24 +00:00
Per PR review suggestion: - `cuda.parallel` - device-wide algorithms/Thrust - `cuda.cooperative` - Cooperative algorithsm/CUB - `cuda.bench` - Benchmarking/NVBench
98 lines
2.8 KiB
Python
98 lines
2.8 KiB
Python
# Copyright 2025 NVIDIA Corporation
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 with the LLVM exception
|
|
# (the "License"); you may not use this file except in compliance with
|
|
# the License.
|
|
#
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://llvm.org/foundation/relicensing/LICENSE.txt
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
import sys
|
|
import time
|
|
|
|
import cuda.bench as bench
|
|
import cuda.cccl.headers as headers
|
|
import cuda.core.experimental as core
|
|
|
|
host_sleep_duration = 0.1
|
|
|
|
|
|
def cpu_only_sleep_bench(state: bench.State) -> None:
|
|
def launcher(launch: bench.Launch):
|
|
time.sleep(host_sleep_duration)
|
|
|
|
state.exec(launcher)
|
|
|
|
|
|
def as_core_Stream(cs: bench.CudaStream) -> core.Stream:
|
|
return core.Stream.from_handle(cs.addressof())
|
|
|
|
|
|
def make_sleep_kernel():
|
|
"""JITs sleep_kernel(seconds)"""
|
|
src = r"""
|
|
#include <cuda/std/cstdint>
|
|
#include <cuda/std/chrono>
|
|
|
|
// Each launched thread just sleeps for `seconds`.
|
|
__global__ void sleep_kernel(double seconds) {
|
|
namespace chrono = ::cuda::std::chrono;
|
|
using hr_clock = chrono::high_resolution_clock;
|
|
|
|
auto duration = static_cast<cuda::std::int64_t>(seconds * 1e9);
|
|
const auto ns = chrono::nanoseconds(duration);
|
|
|
|
const auto start = hr_clock::now();
|
|
const auto finish = start + ns;
|
|
|
|
auto now = hr_clock::now();
|
|
while (now < finish)
|
|
{
|
|
now = hr_clock::now();
|
|
}
|
|
}
|
|
"""
|
|
incl = headers.get_include_paths()
|
|
opts = core.ProgramOptions(include_path=str(incl.libcudacxx))
|
|
prog = core.Program(src, code_type="c++", options=opts)
|
|
mod = prog.compile("cubin", name_expressions=("sleep_kernel",))
|
|
return mod.get_kernel("sleep_kernel")
|
|
|
|
|
|
def mixed_sleep_bench(state: bench.State) -> None:
|
|
sync = state.get_string("Sync")
|
|
sync_flag = sync == "Do sync"
|
|
|
|
gpu_sleep_dur = 225e-3
|
|
krn = make_sleep_kernel()
|
|
launch_config = core.LaunchConfig(grid=1, block=1, shmem_size=0)
|
|
|
|
def launcher(launch: bench.Launch):
|
|
# host overhead
|
|
time.sleep(host_sleep_duration)
|
|
# GPU computation
|
|
s = as_core_Stream(launch.get_stream())
|
|
core.launch(s, launch_config, krn, gpu_sleep_dur)
|
|
|
|
state.exec(launcher, sync=sync_flag)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
# time function only doing work (sleeping) on the host
|
|
# using CPU timer only
|
|
b = bench.register(cpu_only_sleep_bench)
|
|
b.set_is_cpu_only(True)
|
|
|
|
# time the function that does work on both GPU and CPU
|
|
b2 = bench.register(mixed_sleep_bench)
|
|
b2.add_string_axis("Sync", ["Do not sync", "Do sync"])
|
|
|
|
bench.run_all_benchmarks(sys.argv)
|