Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ColossalAI
Commits
46c009db
Unverified
Commit
46c009db
authored
Apr 06, 2023
by
Hakjin Lee
Committed by
GitHub
Apr 05, 2023
Browse files
[format] Run lint on colossalai.engine (#3367)
parent
b9231390
Changes
9
Show whitespace changes
Inline
Side-by-side
Showing
9 changed files
with
32 additions
and
20 deletions
+32
-20
colossalai/engine/gradient_accumulation/__init__.py
colossalai/engine/gradient_accumulation/__init__.py
+11
-4
colossalai/engine/gradient_handler/_base_gradient_handler.py
colossalai/engine/gradient_handler/_base_gradient_handler.py
+1
-1
colossalai/engine/gradient_handler/_data_parallel_gradient_handler.py
...ngine/gradient_handler/_data_parallel_gradient_handler.py
+4
-3
colossalai/engine/gradient_handler/_pipeline_parallel_gradient_handler.py
...e/gradient_handler/_pipeline_parallel_gradient_handler.py
+4
-3
colossalai/engine/gradient_handler/_sequence_parallel_gradient_handler.py
...e/gradient_handler/_sequence_parallel_gradient_handler.py
+4
-3
colossalai/engine/gradient_handler/_zero_gradient_handler.py
colossalai/engine/gradient_handler/_zero_gradient_handler.py
+1
-0
colossalai/engine/schedule/__init__.py
colossalai/engine/schedule/__init__.py
+1
-1
colossalai/engine/schedule/_base_schedule.py
colossalai/engine/schedule/_base_schedule.py
+1
-1
colossalai/engine/schedule/_non_pipeline_schedule.py
colossalai/engine/schedule/_non_pipeline_schedule.py
+5
-4
No files found.
colossalai/engine/gradient_accumulation/__init__.py
View file @
46c009db
from
typing
import
Iterable
,
List
import
torch.nn
as
nn
import
torch.nn
as
nn
from
typing
import
List
from
colossalai.engine
import
BaseGradientHandler
from
typing
import
Iterable
from
torch.optim
import
Optimizer
from
torch.optim
import
Optimizer
from
torch.optim.lr_scheduler
import
_LRScheduler
from
torch.optim.lr_scheduler
import
_LRScheduler
from
._gradient_accumulation
import
GradAccumDataloader
,
GradAccumOptimizer
,
GradAccumLrSchedulerByStep
,
GradAccumGradientHandler
from
colossalai.engine
import
BaseGradientHandler
from
._gradient_accumulation
import
(
GradAccumDataloader
,
GradAccumGradientHandler
,
GradAccumLrSchedulerByStep
,
GradAccumOptimizer
,
)
__all__
=
[
__all__
=
[
'accumulate_gradient'
,
'GradAccumDataloader'
,
'GradAccumOptimizer'
,
'GradAccumLrSchedulerByStep'
,
'accumulate_gradient'
,
'GradAccumDataloader'
,
'GradAccumOptimizer'
,
'GradAccumLrSchedulerByStep'
,
...
...
colossalai/engine/gradient_handler/_base_gradient_handler.py
View file @
46c009db
colossalai/engine/gradient_handler/_data_parallel_gradient_handler.py
View file @
46c009db
from
colossalai.core
import
global_context
as
gpc
from
colossalai.core
import
global_context
as
gpc
from
colossalai.registry
import
GRADIENT_HANDLER
from
colossalai.registry
import
GRADIENT_HANDLER
from
._base_gradient_handler
import
BaseGradientHandler
from
...context.parallel_mode
import
ParallelMode
from
...context.parallel_mode
import
ParallelMode
from
._base_gradient_handler
import
BaseGradientHandler
from
.utils
import
bucket_allreduce
from
.utils
import
bucket_allreduce
...
...
colossalai/engine/gradient_handler/_pipeline_parallel_gradient_handler.py
View file @
46c009db
...
@@ -4,9 +4,10 @@ from collections import defaultdict
...
@@ -4,9 +4,10 @@ from collections import defaultdict
import
torch
import
torch
import
torch.distributed
as
dist
import
torch.distributed
as
dist
from
torch._utils
import
_flatten_dense_tensors
,
_unflatten_dense_tensors
from
colossalai.core
import
global_context
as
gpc
from
colossalai.core
import
global_context
as
gpc
from
colossalai.registry
import
GRADIENT_HANDLER
from
colossalai.registry
import
GRADIENT_HANDLER
from
torch._utils
import
_flatten_dense_tensors
,
_unflatten_dense_tensors
from
._base_gradient_handler
import
BaseGradientHandler
from
._base_gradient_handler
import
BaseGradientHandler
...
...
colossalai/engine/gradient_handler/_sequence_parallel_gradient_handler.py
View file @
46c009db
from
colossalai.core
import
global_context
as
gpc
from
colossalai.core
import
global_context
as
gpc
from
colossalai.registry
import
GRADIENT_HANDLER
from
colossalai.registry
import
GRADIENT_HANDLER
from
._base_gradient_handler
import
BaseGradientHandler
from
...context.parallel_mode
import
ParallelMode
from
...context.parallel_mode
import
ParallelMode
from
._base_gradient_handler
import
BaseGradientHandler
from
.utils
import
bucket_allreduce
from
.utils
import
bucket_allreduce
...
...
colossalai/engine/gradient_handler/_zero_gradient_handler.py
View file @
46c009db
from
colossalai.registry
import
GRADIENT_HANDLER
from
colossalai.registry
import
GRADIENT_HANDLER
from
._base_gradient_handler
import
BaseGradientHandler
from
._base_gradient_handler
import
BaseGradientHandler
...
...
colossalai/engine/schedule/__init__.py
View file @
46c009db
from
._base_schedule
import
BaseSchedule
from
._base_schedule
import
BaseSchedule
from
._pipeline_schedule
import
PipelineSchedule
,
InterleavedPipelineSchedule
,
get_tensor_shape
from
._non_pipeline_schedule
import
NonPipelineSchedule
from
._non_pipeline_schedule
import
NonPipelineSchedule
from
._pipeline_schedule
import
InterleavedPipelineSchedule
,
PipelineSchedule
,
get_tensor_shape
__all__
=
[
'BaseSchedule'
,
'NonPipelineSchedule'
,
'PipelineSchedule'
,
'InterleavedPipelineSchedule'
,
'get_tensor_shape'
]
__all__
=
[
'BaseSchedule'
,
'NonPipelineSchedule'
,
'PipelineSchedule'
,
'InterleavedPipelineSchedule'
,
'get_tensor_shape'
]
colossalai/engine/schedule/_base_schedule.py
View file @
46c009db
...
@@ -2,10 +2,10 @@
...
@@ -2,10 +2,10 @@
# -*- encoding: utf-8 -*-
# -*- encoding: utf-8 -*-
from
abc
import
ABC
,
abstractmethod
from
abc
import
ABC
,
abstractmethod
from
typing
import
Callable
,
Iterable
import
torch
import
torch
from
typing
import
Iterable
,
Callable
from
colossalai.logging
import
get_dist_logger
from
colossalai.logging
import
get_dist_logger
from
colossalai.utils
import
get_current_device
from
colossalai.utils
import
get_current_device
...
...
colossalai/engine/schedule/_non_pipeline_schedule.py
View file @
46c009db
#!/usr/bin/env python
#!/usr/bin/env python
# -*- encoding: utf-8 -*-
# -*- encoding: utf-8 -*-
from
typing
import
Iterable
import
inspect
from
typing
import
Callable
,
Iterable
import
torch
import
torch
import
inspect
from
._base_schedule
import
BaseSchedule
from
colossalai.utils
import
conditional_context
from
colossalai.utils
import
conditional_context
from
typing
import
Callable
from
._base_schedule
import
BaseSchedule
class
NonPipelineSchedule
(
BaseSchedule
):
class
NonPipelineSchedule
(
BaseSchedule
):
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment