mirror of
https://github.com/ikawrakow/ik_llama.cpp.git
synced 2026-04-21 06:59:21 +00:00
* Fix bf16 additions on CUDA arch < Ampere * Prevent using NCCL if graph reduce type is bf16 and arch < AMPERE
* Fix bf16 additions on CUDA arch < Ampere * Prevent using NCCL if graph reduce type is bf16 and arch < AMPERE