Files
tabbyAPI/backends/exllamav3/utils.py
2025-10-09 22:33:53 -04:00

15 lines
388 B
Python

import platform
import torch
from loguru import logger
def exllama_supports_nccl():
if platform.system() == "Windows":
unsupported_message = (
"The NCCL tensor parallel backend is not supported on Windows."
)
logger.warning(unsupported_message)
return False
return torch.cuda.is_available() and torch.distributed.is_nccl_available()