From 3a9cf1f8e5b64e6abaebec256000f881f1490b67 Mon Sep 17 00:00:00 2001 From: layerdiffusion <19834515+lllyasviel@users.noreply.github.com> Date: Sat, 31 Aug 2024 11:07:28 -0700 Subject: [PATCH] Revert partially "use safer codes" --- packages_3rdparty/gguf/quants.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/packages_3rdparty/gguf/quants.py b/packages_3rdparty/gguf/quants.py index ac82f79b..1fac91ff 100644 --- a/packages_3rdparty/gguf/quants.py +++ b/packages_3rdparty/gguf/quants.py @@ -151,7 +151,7 @@ class __Quant(ABC): rows = data.reshape((-1, data.shape[-1])).view(torch.uint8) n_blocks = rows.numel() // cls.type_size blocks = rows.reshape((n_blocks, cls.type_size)) - parameter.data = blocks.clone(memory_format=torch.contiguous_format) + parameter.data = blocks.contiguous() cls.bake_inner(parameter) parameter.baked = True return @@ -312,7 +312,7 @@ class Q4_0(__Quant, qtype=GGMLQuantizationType.Q4_0): d, x = quick_split(blocks, [2]) d = d.view(torch.float16).to(parameter.computation_dtype).view(torch.uint8) x = change_4bits_order(x).view(torch.uint8) - parameter.data = torch.cat([d, x], dim=-1).clone(memory_format=torch.contiguous_format) + parameter.data = torch.cat([d, x], dim=-1).contiguous() return @classmethod @@ -389,7 +389,7 @@ class Q4_1(__Quant, qtype=GGMLQuantizationType.Q4_1): m = m.view(torch.float16).to(parameter.computation_dtype).view(torch.uint8) qs = change_4bits_order(qs).view(torch.uint8) - parameter.data = torch.cat([d, m, qs], dim=-1).clone(memory_format=torch.contiguous_format) + parameter.data = torch.cat([d, m, qs], dim=-1).contiguous() return @@ -601,7 +601,7 @@ class Q8_0(__Quant, qtype=GGMLQuantizationType.Q8_0): d, x = quick_split(blocks, [2]) x = x.view(torch.int8) d = d.view(torch.float16).to(parameter.computation_dtype).view(torch.int8) - parameter.data = torch.cat([d, x], dim=-1).clone(memory_format=torch.contiguous_format) + parameter.data = torch.cat([d, x], dim=-1).contiguous() return @classmethod @@ -808,7 +808,7 @@ class Q4_K(__Quant, qtype=GGMLQuantizationType.Q4_K): dm = dm.view(torch.uint8).reshape((n_blocks, -1)) qs = qs.view(torch.uint8) - parameter.data = torch.cat([d, dm, qs], dim=-1).clone(memory_format=torch.contiguous_format) + parameter.data = torch.cat([d, dm, qs], dim=-1).contiguous() return @classmethod