mirror of
https://github.com/turboderp-org/exllamav3.git
synced 2026-04-20 14:29:51 +00:00
ppl_transformers.py: Explicitly make bfloat16 the default dtype
This commit is contained in:
@@ -48,7 +48,7 @@ def main(args):
|
||||
model = AutoModelForCausalLM.from_pretrained(
|
||||
args.model_dir,
|
||||
device_map = "auto",
|
||||
torch_dtype = torch.half if args.tight else torch.float if args.fp32 else None,
|
||||
torch_dtype = torch.half if args.tight else torch.float if args.fp32 else torch.bfloat16,
|
||||
)
|
||||
if args.tight:
|
||||
free_mem()
|
||||
|
||||
Reference in New Issue
Block a user