Commit 539bad24 authored by lcskrishna's avatar lcskrishna
Browse files

cleanup of extensions

parent 9b4c68c7
...@@ -202,20 +202,6 @@ if "--cuda_ext" in sys.argv: ...@@ -202,20 +202,6 @@ if "--cuda_ext" in sys.argv:
'csrc/hip/multi_tensor_novograd.hip', 'csrc/hip/multi_tensor_novograd.hip',
'csrc/hip/multi_tensor_lamb.hip', 'csrc/hip/multi_tensor_lamb.hip',
] ]
#ext_modules.append(
# CUDAExtension(name='amp_C',
# sources=['csrc/amp_C_frontend.cpp',
# 'csrc/hip/multi_tensor_sgd_kernel.hip',
# 'csrc/hip/multi_tensor_scale_kernel.hip',
# 'csrc/hip/multi_tensor_axpby_kernel.hip',
# 'csrc/hip/multi_tensor_l2norm_kernel.hip',
# 'csrc/hip/multi_tensor_lamb_stage_1.hip',
# 'csrc/hip/multi_tensor_lamb_stage_2.hip',
# 'csrc/hip/multi_tensor_adam.hip',
# 'csrc/hip/multi_tensor_adagrad.hip',
# 'csrc/hip/multi_tensor_novograd.hip',
# 'csrc/hip/multi_tensor_lamb.hip'],
# extra_compile_args=['-O3'] + version_dependent_macros))
ext_modules.append( ext_modules.append(
CUDAExtension(name='amp_C', CUDAExtension(name='amp_C',
sources=multi_tensor_sources_v1_8 if torch.__version__ >= '1.8' else multi_tensor_sources_other, sources=multi_tensor_sources_v1_8 if torch.__version__ >= '1.8' else multi_tensor_sources_other,
...@@ -236,12 +222,6 @@ if "--cuda_ext" in sys.argv: ...@@ -236,12 +222,6 @@ if "--cuda_ext" in sys.argv:
CUDAExtension(name='syncbn', CUDAExtension(name='syncbn',
sources=syncbn_sources_v1_8 if torch.__version__ >= '1.8' else syncbn_sources_other, sources=syncbn_sources_v1_8 if torch.__version__ >= '1.8' else syncbn_sources_other,
extra_compile_args=['-O3'] + version_dependent_macros)) extra_compile_args=['-O3'] + version_dependent_macros))
#ext_modules.append(
# CUDAExtension(name='syncbn',
# sources=['csrc/syncbn.cpp',
# 'csrc/hip/welford.hip'],
# extra_compile_args=['-O3'] + version_dependent_macros))
if not is_rocm_pytorch: if not is_rocm_pytorch:
ext_modules.append( ext_modules.append(
...@@ -260,12 +240,6 @@ if "--cuda_ext" in sys.argv: ...@@ -260,12 +240,6 @@ if "--cuda_ext" in sys.argv:
CUDAExtension(name='fused_layer_norm_cuda', CUDAExtension(name='fused_layer_norm_cuda',
sources = layer_norm_sources_v1_8 if torch.__version__ >= '1.8' else layer_norm_sources_other, sources = layer_norm_sources_v1_8 if torch.__version__ >= '1.8' else layer_norm_sources_other,
extra_compile_args=['-O3'] + version_dependent_macros)) extra_compile_args=['-O3'] + version_dependent_macros))
#ext_modules.append(
# CUDAExtension(name='fused_layer_norm_cuda',
# sources=['csrc/layer_norm_cuda.cpp',
# 'csrc/hip/layer_norm_hip_kernel.hip'],
# extra_compile_args={'cxx' : ['-O3'] + version_dependent_macros,
# 'nvcc' : []}))
if not is_rocm_pytorch: if not is_rocm_pytorch:
ext_modules.append( ext_modules.append(
...@@ -282,12 +256,6 @@ if "--cuda_ext" in sys.argv: ...@@ -282,12 +256,6 @@ if "--cuda_ext" in sys.argv:
CUDAExtension(name='mlp_cuda', CUDAExtension(name='mlp_cuda',
sources = mlp_sources_v1_8 if torch.__version__ >= '1.8' else mlp_sources_other, sources = mlp_sources_v1_8 if torch.__version__ >= '1.8' else mlp_sources_other,
extra_compile_args=['-O3'] + version_dependent_macros)) extra_compile_args=['-O3'] + version_dependent_macros))
#ext_modules.append(
# CUDAExtension(name='mlp_cuda',
# sources=['csrc/mlp.cpp',
# 'csrc/hip/mlp_hip.hip'],
# extra_compile_args={'cxx' : ['-O3'] + version_dependent_macros,
# 'nvcc' : []}))
if "--bnp" in sys.argv: if "--bnp" in sys.argv:
from torch.utils.cpp_extension import CUDAExtension from torch.utils.cpp_extension import CUDAExtension
...@@ -341,12 +309,6 @@ if "--xentropy" in sys.argv: ...@@ -341,12 +309,6 @@ if "--xentropy" in sys.argv:
sources = xentropy_sources_v1_8 if torch.__version__ >= '1.8' else xentropy_sources_other, sources = xentropy_sources_v1_8 if torch.__version__ >= '1.8' else xentropy_sources_other,
include_dirs=[os.path.join(this_dir, 'csrc') if torch.__version__ >= '1.8' else os.path.join(this_dir, 'csrc/hip')], include_dirs=[os.path.join(this_dir, 'csrc') if torch.__version__ >= '1.8' else os.path.join(this_dir, 'csrc/hip')],
extra_compile_args=['-O3'] + version_dependent_macros)) extra_compile_args=['-O3'] + version_dependent_macros))
#ext_modules.append(
# CUDAExtension(name='xentropy_cuda',
# sources=['apex/contrib/csrc/xentropy/interface.cpp',
# 'apex/contrib/csrc/xentropy/hip/xentropy_kernel.hip'],
# include_dirs=[os.path.join(this_dir, 'csrc/hip')],
# extra_compile_args=['-O3'] + version_dependent_macros))
if "--deprecated_fused_adam" in sys.argv: if "--deprecated_fused_adam" in sys.argv:
...@@ -383,12 +345,6 @@ if "--deprecated_fused_adam" in sys.argv: ...@@ -383,12 +345,6 @@ if "--deprecated_fused_adam" in sys.argv:
sources = fused_adam_sources_v1_8 if torch.__version__ >= '1.8' else fused_adam_sources_other, sources = fused_adam_sources_v1_8 if torch.__version__ >= '1.8' else fused_adam_sources_other,
include_dirs=[os.path.join(this_dir, 'csrc') if torch.__version__ >= '1.8' else os.path.join(this_dir, 'csrc/hip')], include_dirs=[os.path.join(this_dir, 'csrc') if torch.__version__ >= '1.8' else os.path.join(this_dir, 'csrc/hip')],
extra_compile_args=['-O3'] + version_dependent_macros)) extra_compile_args=['-O3'] + version_dependent_macros))
#ext_modules.append(
# CUDAExtension(name='fused_adam_cuda',
# sources=['apex/contrib/csrc/optimizers/fused_adam_cuda.cpp',
# 'apex/contrib/csrc/optimizers/hip/fused_adam_hip_kernel.hip'],
# include_dirs=[os.path.join(this_dir, 'csrc/hip')],
# extra_compile_args=['-O3'] + version_dependent_macros))
if "--deprecated_fused_lamb" in sys.argv: if "--deprecated_fused_lamb" in sys.argv:
from torch.utils.cpp_extension import CUDAExtension from torch.utils.cpp_extension import CUDAExtension
...@@ -425,13 +381,6 @@ if "--deprecated_fused_lamb" in sys.argv: ...@@ -425,13 +381,6 @@ if "--deprecated_fused_lamb" in sys.argv:
sources = fused_lamb_sources_v1_8 if torch.__version__ >= '1.8' else fused_lamb_sources_other, sources = fused_lamb_sources_v1_8 if torch.__version__ >= '1.8' else fused_lamb_sources_other,
include_dirs = [os.path.join(this_dir, 'csrc') if torch.__version__ >= '1.8' else os.path.join(this_dir, 'csrc/hip')], include_dirs = [os.path.join(this_dir, 'csrc') if torch.__version__ >= '1.8' else os.path.join(this_dir, 'csrc/hip')],
extra_compile_args=['-O3'] + version_dependent_macros)) extra_compile_args=['-O3'] + version_dependent_macros))
#ext_modules.append(
# CUDAExtension(name='fused_lamb_cuda',
# sources=['apex/contrib/csrc/optimizers/fused_lamb_cuda.cpp',
# 'apex/contrib/csrc/optimizers/hip/fused_lamb_hip_kernel.hip',
# 'csrc/hip/multi_tensor_l2norm_kernel.hip'],
# include_dirs=[os.path.join(this_dir, 'csrc/hip')],
# extra_compile_args=['-O3'] + version_dependent_macros))
# Check, if ATen/CUDAGenerator.h is found, otherwise use the new ATen/CUDAGeneratorImpl.h, due to breaking change in https://github.com/pytorch/pytorch/pull/36026 # Check, if ATen/CUDAGenerator.h is found, otherwise use the new ATen/CUDAGeneratorImpl.h, due to breaking change in https://github.com/pytorch/pytorch/pull/36026
generator_flag = [] generator_flag = []
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment