Commit d76078ef authored by yangql's avatar yangql
Browse files

initing develop

parent a2630e0f
...@@ -9,16 +9,16 @@ import transformers ...@@ -9,16 +9,16 @@ import transformers
logger = getLogger(__name__) logger = getLogger(__name__)
try: # try:
import autogptq_cuda_64 # import autogptq_cuda_64
import autogptq_cuda_256 # import autogptq_cuda_256
_autogptq_cuda_available = True # _autogptq_cuda_available = True
except ImportError: # except ImportError:
logger.warning("CUDA extension not installed.") # logger.warning("CUDA extension not installed.")
autogptq_cuda_256 = None autogptq_cuda_256 = None
autogptq_cuda_64 = None autogptq_cuda_64 = None
_autogptq_cuda_available = False _autogptq_cuda_available = False
class QuantLinear(nn.Module): class QuantLinear(nn.Module):
......
...@@ -8,16 +8,16 @@ import transformers ...@@ -8,16 +8,16 @@ import transformers
logger = getLogger(__name__) logger = getLogger(__name__)
try: # try:
import autogptq_cuda_64 # import autogptq_cuda_64
import autogptq_cuda_256 # import autogptq_cuda_256
_autogptq_cuda_available = True # _autogptq_cuda_available = True
except ImportError: # except ImportError:
logger.warning("CUDA extension not installed.") # logger.warning("CUDA extension not installed.")
autogptq_cuda_256 = None autogptq_cuda_256 = None
autogptq_cuda_64 = None autogptq_cuda_64 = None
_autogptq_cuda_available = False _autogptq_cuda_available = False
class QuantLinear(nn.Module): class QuantLinear(nn.Module):
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment