Clean up imports

This commit is contained in:
turboderp
2023-09-11 07:31:43 +02:00
parent 27bccf02d3
commit c5c90a8b4b
10 changed files with 3 additions and 15 deletions

View File

@@ -2,8 +2,6 @@ import torch
from torch import nn
import torch.nn.functional as F
import math
import exllamav2.ext
from exllamav2 import ext
from exllamav2.ext import exllamav2_ext as ext_c, none_tensor

View File

@@ -5,7 +5,7 @@ from conversion.qparams import QParams, qparams_options, qparams_headoptions
from conversion.adaptivegptq import AdaptiveGPTQ
import torch
from torch import nn
import os, sys, time, math, json
import os, time, math, json
import torch.nn.functional as F
import gc

View File

@@ -2,7 +2,7 @@
import sys, os
sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
from exllamav2 import(
from exllamav2 import (
ExLlamaV2,
ExLlamaV2Config,
ExLlamaV2Cache,

View File

@@ -8,8 +8,6 @@ from exllamav2.cache import ExLlamaV2Cache
import math
from exllamav2 import ext
from exllamav2.ext import exllamav2_ext as ext_c, none_tensor
import gc
import sys
# from flash_attn import flash_attn_func
# import xformers.ops as xops
# from exllamav2.util import list_live_tensors, set_snapshot, diff_snapshot, print_vram_usage_peak

View File

@@ -1,6 +1,6 @@
import torch
from safetensors import safe_open
import sys, os, glob, json
import os, glob, json
class ExLlamaV2Config:

View File

@@ -1,7 +1,6 @@
import torch
from exllamav2.module import ExLlamaV2Module
from torch import nn
import math
from exllamav2 import ext
from exllamav2.ext import exllamav2_ext as ext_c, none_tensor
from safetensors import safe_open

View File

@@ -1,10 +1,8 @@
import torch
from torch import nn
import torch.nn.functional as F
from exllamav2.module import ExLlamaV2Module
from exllamav2.rmsnorm import ExLlamaV2RMSNorm
from exllamav2.linear import ExLlamaV2Linear
from exllamav2 import ext
from exllamav2.ext import exllamav2_ext as ext_c, none_tensor
class ExLlamaV2MLP(ExLlamaV2Module):

View File

@@ -6,9 +6,6 @@ if sys.version_info < min_version:
print("")
import torch
from torch import nn
import torch.nn.functional as F
from safetensors import safe_open
import math
from exllamav2.config import ExLlamaV2Config
from exllamav2.cache import ExLlamaV2Cache

View File

@@ -4,7 +4,6 @@ import argparse, sys, os, glob
from exllamav2 import(
ExLlamaV2,
ExLlamaV2Config,
ExLlamaV2Cache,
ExLlamaV2Tokenizer
)

View File

@@ -1,7 +1,6 @@
import torch
from torch import nn
from exllamav2.module import ExLlamaV2Module
from exllamav2 import ext
from exllamav2.ext import exllamav2_ext as ext_c, none_tensor
class ExLlamaV2RMSNorm(ExLlamaV2Module):