Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
apex
Commits
539bad24
Commit
539bad24
authored
Dec 10, 2020
by
lcskrishna
Browse files
cleanup of extensions
parent
9b4c68c7
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
0 additions
and
51 deletions
+0
-51
setup.py
setup.py
+0
-51
No files found.
setup.py
View file @
539bad24
...
...
@@ -202,20 +202,6 @@ if "--cuda_ext" in sys.argv:
'csrc/hip/multi_tensor_novograd.hip'
,
'csrc/hip/multi_tensor_lamb.hip'
,
]
#ext_modules.append(
# CUDAExtension(name='amp_C',
# sources=['csrc/amp_C_frontend.cpp',
# 'csrc/hip/multi_tensor_sgd_kernel.hip',
# 'csrc/hip/multi_tensor_scale_kernel.hip',
# 'csrc/hip/multi_tensor_axpby_kernel.hip',
# 'csrc/hip/multi_tensor_l2norm_kernel.hip',
# 'csrc/hip/multi_tensor_lamb_stage_1.hip',
# 'csrc/hip/multi_tensor_lamb_stage_2.hip',
# 'csrc/hip/multi_tensor_adam.hip',
# 'csrc/hip/multi_tensor_adagrad.hip',
# 'csrc/hip/multi_tensor_novograd.hip',
# 'csrc/hip/multi_tensor_lamb.hip'],
# extra_compile_args=['-O3'] + version_dependent_macros))
ext_modules
.
append
(
CUDAExtension
(
name
=
'amp_C'
,
sources
=
multi_tensor_sources_v1_8
if
torch
.
__version__
>=
'1.8'
else
multi_tensor_sources_other
,
...
...
@@ -236,12 +222,6 @@ if "--cuda_ext" in sys.argv:
CUDAExtension
(
name
=
'syncbn'
,
sources
=
syncbn_sources_v1_8
if
torch
.
__version__
>=
'1.8'
else
syncbn_sources_other
,
extra_compile_args
=
[
'-O3'
]
+
version_dependent_macros
))
#ext_modules.append(
# CUDAExtension(name='syncbn',
# sources=['csrc/syncbn.cpp',
# 'csrc/hip/welford.hip'],
# extra_compile_args=['-O3'] + version_dependent_macros))
if
not
is_rocm_pytorch
:
ext_modules
.
append
(
...
...
@@ -260,12 +240,6 @@ if "--cuda_ext" in sys.argv:
CUDAExtension
(
name
=
'fused_layer_norm_cuda'
,
sources
=
layer_norm_sources_v1_8
if
torch
.
__version__
>=
'1.8'
else
layer_norm_sources_other
,
extra_compile_args
=
[
'-O3'
]
+
version_dependent_macros
))
#ext_modules.append(
# CUDAExtension(name='fused_layer_norm_cuda',
# sources=['csrc/layer_norm_cuda.cpp',
# 'csrc/hip/layer_norm_hip_kernel.hip'],
# extra_compile_args={'cxx' : ['-O3'] + version_dependent_macros,
# 'nvcc' : []}))
if
not
is_rocm_pytorch
:
ext_modules
.
append
(
...
...
@@ -282,12 +256,6 @@ if "--cuda_ext" in sys.argv:
CUDAExtension
(
name
=
'mlp_cuda'
,
sources
=
mlp_sources_v1_8
if
torch
.
__version__
>=
'1.8'
else
mlp_sources_other
,
extra_compile_args
=
[
'-O3'
]
+
version_dependent_macros
))
#ext_modules.append(
# CUDAExtension(name='mlp_cuda',
# sources=['csrc/mlp.cpp',
# 'csrc/hip/mlp_hip.hip'],
# extra_compile_args={'cxx' : ['-O3'] + version_dependent_macros,
# 'nvcc' : []}))
if
"--bnp"
in
sys
.
argv
:
from
torch.utils.cpp_extension
import
CUDAExtension
...
...
@@ -341,12 +309,6 @@ if "--xentropy" in sys.argv:
sources
=
xentropy_sources_v1_8
if
torch
.
__version__
>=
'1.8'
else
xentropy_sources_other
,
include_dirs
=
[
os
.
path
.
join
(
this_dir
,
'csrc'
)
if
torch
.
__version__
>=
'1.8'
else
os
.
path
.
join
(
this_dir
,
'csrc/hip'
)],
extra_compile_args
=
[
'-O3'
]
+
version_dependent_macros
))
#ext_modules.append(
# CUDAExtension(name='xentropy_cuda',
# sources=['apex/contrib/csrc/xentropy/interface.cpp',
# 'apex/contrib/csrc/xentropy/hip/xentropy_kernel.hip'],
# include_dirs=[os.path.join(this_dir, 'csrc/hip')],
# extra_compile_args=['-O3'] + version_dependent_macros))
if
"--deprecated_fused_adam"
in
sys
.
argv
:
...
...
@@ -383,12 +345,6 @@ if "--deprecated_fused_adam" in sys.argv:
sources
=
fused_adam_sources_v1_8
if
torch
.
__version__
>=
'1.8'
else
fused_adam_sources_other
,
include_dirs
=
[
os
.
path
.
join
(
this_dir
,
'csrc'
)
if
torch
.
__version__
>=
'1.8'
else
os
.
path
.
join
(
this_dir
,
'csrc/hip'
)],
extra_compile_args
=
[
'-O3'
]
+
version_dependent_macros
))
#ext_modules.append(
# CUDAExtension(name='fused_adam_cuda',
# sources=['apex/contrib/csrc/optimizers/fused_adam_cuda.cpp',
# 'apex/contrib/csrc/optimizers/hip/fused_adam_hip_kernel.hip'],
# include_dirs=[os.path.join(this_dir, 'csrc/hip')],
# extra_compile_args=['-O3'] + version_dependent_macros))
if
"--deprecated_fused_lamb"
in
sys
.
argv
:
from
torch.utils.cpp_extension
import
CUDAExtension
...
...
@@ -425,13 +381,6 @@ if "--deprecated_fused_lamb" in sys.argv:
sources
=
fused_lamb_sources_v1_8
if
torch
.
__version__
>=
'1.8'
else
fused_lamb_sources_other
,
include_dirs
=
[
os
.
path
.
join
(
this_dir
,
'csrc'
)
if
torch
.
__version__
>=
'1.8'
else
os
.
path
.
join
(
this_dir
,
'csrc/hip'
)],
extra_compile_args
=
[
'-O3'
]
+
version_dependent_macros
))
#ext_modules.append(
# CUDAExtension(name='fused_lamb_cuda',
# sources=['apex/contrib/csrc/optimizers/fused_lamb_cuda.cpp',
# 'apex/contrib/csrc/optimizers/hip/fused_lamb_hip_kernel.hip',
# 'csrc/hip/multi_tensor_l2norm_kernel.hip'],
# include_dirs=[os.path.join(this_dir, 'csrc/hip')],
# extra_compile_args=['-O3'] + version_dependent_macros))
# Check, if ATen/CUDAGenerator.h is found, otherwise use the new ATen/CUDAGeneratorImpl.h, due to breaking change in https://github.com/pytorch/pytorch/pull/36026
generator_flag
=
[]
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment