Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
fairscale
Commits
8f8f8ef9
Unverified
Commit
8f8f8ef9
authored
Sep 23, 2022
by
Min Xu
Committed by
GitHub
Sep 23, 2022
Browse files
[chore] move fair_dev into fairscale (#1078)
Co-authored-by:
Min Xu
<
min.xu.public@gmail.com
>
parent
bfd57ff3
Changes
63
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
33 additions
and
21 deletions
+33
-21
tests/nn/data_parallel/test_fsdp_freezing_weights.py
tests/nn/data_parallel/test_fsdp_freezing_weights.py
+1
-1
tests/nn/data_parallel/test_fsdp_grad_acc.py
tests/nn/data_parallel/test_fsdp_grad_acc.py
+1
-1
tests/nn/data_parallel/test_fsdp_hf_transformer_eval.py
tests/nn/data_parallel/test_fsdp_hf_transformer_eval.py
+1
-1
tests/nn/data_parallel/test_fsdp_input.py
tests/nn/data_parallel/test_fsdp_input.py
+1
-1
tests/nn/data_parallel/test_fsdp_memory.py
tests/nn/data_parallel/test_fsdp_memory.py
+1
-1
tests/nn/data_parallel/test_fsdp_metadata.py
tests/nn/data_parallel/test_fsdp_metadata.py
+1
-1
tests/nn/data_parallel/test_fsdp_multiple_forward.py
tests/nn/data_parallel/test_fsdp_multiple_forward.py
+1
-1
tests/nn/data_parallel/test_fsdp_multiple_forward_checkpoint.py
...nn/data_parallel/test_fsdp_multiple_forward_checkpoint.py
+1
-1
tests/nn/data_parallel/test_fsdp_multiple_wrapping.py
tests/nn/data_parallel/test_fsdp_multiple_wrapping.py
+1
-1
tests/nn/data_parallel/test_fsdp_offload.py
tests/nn/data_parallel/test_fsdp_offload.py
+1
-1
tests/nn/data_parallel/test_fsdp_optimizer_utils.py
tests/nn/data_parallel/test_fsdp_optimizer_utils.py
+1
-1
tests/nn/data_parallel/test_fsdp_overlap.py
tests/nn/data_parallel/test_fsdp_overlap.py
+7
-1
tests/nn/data_parallel/test_fsdp_pre_backward_hook.py
tests/nn/data_parallel/test_fsdp_pre_backward_hook.py
+1
-1
tests/nn/data_parallel/test_fsdp_regnet.py
tests/nn/data_parallel/test_fsdp_regnet.py
+1
-1
tests/nn/data_parallel/test_fsdp_shared_weights.py
tests/nn/data_parallel/test_fsdp_shared_weights.py
+7
-1
tests/nn/data_parallel/test_fsdp_shared_weights_mevo.py
tests/nn/data_parallel/test_fsdp_shared_weights_mevo.py
+2
-2
tests/nn/data_parallel/test_fsdp_state_dict.py
tests/nn/data_parallel/test_fsdp_state_dict.py
+1
-1
tests/nn/data_parallel/test_fsdp_uneven.py
tests/nn/data_parallel/test_fsdp_uneven.py
+1
-1
tests/nn/data_parallel/test_fsdp_with_checkpoint_wrapper.py
tests/nn/data_parallel/test_fsdp_with_checkpoint_wrapper.py
+1
-1
tests/nn/data_parallel/test_sharded_ddp_features.py
tests/nn/data_parallel/test_sharded_ddp_features.py
+1
-1
No files found.
tests/nn/data_parallel/test_fsdp_freezing_weights.py
View file @
8f8f8ef9
...
@@ -21,7 +21,7 @@ import torch.nn as nn
...
@@ -21,7 +21,7 @@ import torch.nn as nn
from
torch.nn.parallel
import
DistributedDataParallel
from
torch.nn.parallel
import
DistributedDataParallel
import
torch.optim
as
optim
import
torch.optim
as
optim
from
fair_dev.testing.testing
import
dist_init
,
objects_are_equal
,
rmf
,
skip_if_single_gpu
,
teardown
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
objects_are_equal
,
rmf
,
skip_if_single_gpu
,
teardown
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
...
...
tests/nn/data_parallel/test_fsdp_grad_acc.py
View file @
8f8f8ef9
...
@@ -12,7 +12,7 @@ from unittest.mock import patch
...
@@ -12,7 +12,7 @@ from unittest.mock import patch
from
parameterized
import
parameterized
from
parameterized
import
parameterized
import
torch
import
torch
from
fair_dev.testing.testing
import
DummyProcessGroup
,
make_cudnn_deterministic
,
objects_are_equal
from
fairscale.
fair_dev.testing.testing
import
DummyProcessGroup
,
make_cudnn_deterministic
,
objects_are_equal
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
from
.test_fsdp
import
DistributedTest
,
NestedWrappedModule
,
rename_test
,
spawn_and_init
from
.test_fsdp
import
DistributedTest
,
NestedWrappedModule
,
rename_test
,
spawn_and_init
...
...
tests/nn/data_parallel/test_fsdp_hf_transformer_eval.py
View file @
8f8f8ef9
...
@@ -6,7 +6,7 @@ import unittest
...
@@ -6,7 +6,7 @@ import unittest
import
torch
import
torch
from
torch
import
nn
from
torch
import
nn
from
fair_dev.testing.testing
import
dist_init
from
fairscale.
fair_dev.testing.testing
import
dist_init
from
fairscale.nn
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn
import
auto_wrap
,
enable_wrap
from
fairscale.nn
import
auto_wrap
,
enable_wrap
...
...
tests/nn/data_parallel/test_fsdp_input.py
View file @
8f8f8ef9
...
@@ -16,7 +16,7 @@ import torch
...
@@ -16,7 +16,7 @@ import torch
from
torch.nn
import
Linear
,
Module
from
torch.nn
import
Linear
,
Module
from
torch.optim
import
SGD
from
torch.optim
import
SGD
from
fair_dev.testing.testing
import
dist_init
,
rmf
,
skip_if_no_cuda
,
teardown
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
rmf
,
skip_if_no_cuda
,
teardown
from
fairscale.internal
import
torch_version
from
fairscale.internal
import
torch_version
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
TrainingState
from
fairscale.nn.data_parallel
import
TrainingState
...
...
tests/nn/data_parallel/test_fsdp_memory.py
View file @
8f8f8ef9
...
@@ -18,7 +18,7 @@ import torch.nn as nn
...
@@ -18,7 +18,7 @@ import torch.nn as nn
from
torch.nn.parallel
import
DistributedDataParallel
from
torch.nn.parallel
import
DistributedDataParallel
import
torch.optim
as
optim
import
torch.optim
as
optim
from
fair_dev.testing.testing
import
dist_init
,
dump_all_tensors
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
dump_all_tensors
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
from
fairscale.internal
import
torch_version
from
fairscale.internal
import
torch_version
from
fairscale.internal.parallel
import
get_process_group_cached
from
fairscale.internal.parallel
import
get_process_group_cached
from
fairscale.nn
import
checkpoint_wrapper
from
fairscale.nn
import
checkpoint_wrapper
...
...
tests/nn/data_parallel/test_fsdp_metadata.py
View file @
8f8f8ef9
...
@@ -14,7 +14,7 @@ import torch.multiprocessing as mp
...
@@ -14,7 +14,7 @@ import torch.multiprocessing as mp
import
torch.nn
as
nn
import
torch.nn
as
nn
from
torch.optim
import
Adam
from
torch.optim
import
Adam
from
fair_dev.testing.testing
import
in_temporary_directory
,
skip_if_single_gpu
,
temp_files_ctx
from
fairscale.
fair_dev.testing.testing
import
in_temporary_directory
,
skip_if_single_gpu
,
temp_files_ctx
from
fairscale.nn
import
FullyShardedDataParallel
from
fairscale.nn
import
FullyShardedDataParallel
from
tests.nn.data_parallel.test_fsdp
import
DistributedTest
,
MixtureOfExperts
,
rename_test
,
spawn_and_init
from
tests.nn.data_parallel.test_fsdp
import
DistributedTest
,
MixtureOfExperts
,
rename_test
,
spawn_and_init
...
...
tests/nn/data_parallel/test_fsdp_multiple_forward.py
View file @
8f8f8ef9
...
@@ -17,7 +17,7 @@ import torch.multiprocessing as mp
...
@@ -17,7 +17,7 @@ import torch.multiprocessing as mp
from
torch.nn
import
Linear
,
Module
from
torch.nn
import
Linear
,
Module
from
torch.optim
import
SGD
from
torch.optim
import
SGD
from
fair_dev.testing.testing
import
dist_init
,
skip_if_single_gpu
,
teardown
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
skip_if_single_gpu
,
teardown
from
fairscale.internal
import
torch_version
from
fairscale.internal
import
torch_version
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
TrainingState
from
fairscale.nn.data_parallel
import
TrainingState
...
...
tests/nn/data_parallel/test_fsdp_multiple_forward_checkpoint.py
View file @
8f8f8ef9
...
@@ -20,7 +20,7 @@ import torch.nn as nn
...
@@ -20,7 +20,7 @@ import torch.nn as nn
from
torch.nn.parallel
import
DistributedDataParallel
from
torch.nn.parallel
import
DistributedDataParallel
import
torch.optim
as
optim
import
torch.optim
as
optim
from
fair_dev.testing.testing
import
dist_init
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
from
fairscale.internal
import
torch_version
from
fairscale.internal
import
torch_version
from
fairscale.nn
import
checkpoint_wrapper
from
fairscale.nn
import
checkpoint_wrapper
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
...
...
tests/nn/data_parallel/test_fsdp_multiple_wrapping.py
View file @
8f8f8ef9
...
@@ -17,7 +17,7 @@ import torch.multiprocessing as mp
...
@@ -17,7 +17,7 @@ import torch.multiprocessing as mp
from
torch.nn
import
Linear
,
Module
,
Sequential
from
torch.nn
import
Linear
,
Module
,
Sequential
from
torch.optim
import
SGD
from
torch.optim
import
SGD
from
fair_dev.testing.testing
import
dist_init
,
skip_if_no_cuda
,
teardown
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
skip_if_no_cuda
,
teardown
from
fairscale.internal
import
torch_version
from
fairscale.internal
import
torch_version
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
TrainingState
from
fairscale.nn.data_parallel
import
TrainingState
...
...
tests/nn/data_parallel/test_fsdp_offload.py
View file @
8f8f8ef9
...
@@ -25,7 +25,7 @@ except ImportError as ie:
...
@@ -25,7 +25,7 @@ except ImportError as ie:
pytestmark
=
pytest
.
mark
.
skipif
(
True
,
reason
=
ie
.
msg
)
pytestmark
=
pytest
.
mark
.
skipif
(
True
,
reason
=
ie
.
msg
)
pass
pass
from
fair_dev.testing.testing
import
dist_init
,
spawn_for_all_world_sizes
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
spawn_for_all_world_sizes
from
fairscale.nn.checkpoint.checkpoint_activations
import
checkpoint_wrapper
from
fairscale.nn.checkpoint.checkpoint_activations
import
checkpoint_wrapper
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
,
OffloadConfig
,
TrainingState
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
,
OffloadConfig
,
TrainingState
...
...
tests/nn/data_parallel/test_fsdp_optimizer_utils.py
View file @
8f8f8ef9
...
@@ -12,7 +12,7 @@ import torch
...
@@ -12,7 +12,7 @@ import torch
from
torch
import
nn
from
torch
import
nn
from
torch.optim
import
SGD
,
Adadelta
,
Adam
# type: ignore
from
torch.optim
import
SGD
,
Adadelta
,
Adam
# type: ignore
from
fair_dev.testing.testing
import
dist_init
,
objects_are_equal
,
spawn_for_all_world_sizes
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
objects_are_equal
,
spawn_for_all_world_sizes
from
fairscale.internal.params
import
recursive_copy_to_device
from
fairscale.internal.params
import
recursive_copy_to_device
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
,
get_fsdp_instances
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
,
get_fsdp_instances
from
fairscale.nn.data_parallel.fsdp_optim_utils
import
is_singleton_tensor
from
fairscale.nn.data_parallel.fsdp_optim_utils
import
is_singleton_tensor
...
...
tests/nn/data_parallel/test_fsdp_overlap.py
View file @
8f8f8ef9
...
@@ -19,7 +19,13 @@ from torch.cuda import Event
...
@@ -19,7 +19,13 @@ from torch.cuda import Event
import
torch.multiprocessing
as
mp
import
torch.multiprocessing
as
mp
import
torch.nn
as
nn
import
torch.nn
as
nn
from
fair_dev.testing.testing
import
dist_init
,
get_cycles_per_ms
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
from
fairscale.fair_dev.testing.testing
import
(
dist_init
,
get_cycles_per_ms
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
,
)
from
fairscale.internal
import
torch_version
from
fairscale.internal
import
torch_version
from
fairscale.nn
import
enable_wrap
,
wrap
from
fairscale.nn
import
enable_wrap
,
wrap
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
...
...
tests/nn/data_parallel/test_fsdp_pre_backward_hook.py
View file @
8f8f8ef9
...
@@ -13,7 +13,7 @@ import pytest
...
@@ -13,7 +13,7 @@ import pytest
import
torch
import
torch
from
torch.nn
import
Linear
,
Module
from
torch.nn
import
Linear
,
Module
from
fair_dev.testing.testing
import
dist_init
,
skip_if_no_cuda
,
teardown
,
temp_files_ctx
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
skip_if_no_cuda
,
teardown
,
temp_files_ctx
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
...
...
tests/nn/data_parallel/test_fsdp_regnet.py
View file @
8f8f8ef9
...
@@ -33,7 +33,7 @@ from torch.nn import (
...
@@ -33,7 +33,7 @@ from torch.nn import (
from
torch.nn.parallel
import
DistributedDataParallel
as
DDP
from
torch.nn.parallel
import
DistributedDataParallel
as
DDP
from
torch.optim
import
SGD
from
torch.optim
import
SGD
from
fair_dev.testing.testing
import
(
from
fairscale.
fair_dev.testing.testing
import
(
dist_init
,
dist_init
,
objects_are_equal
,
objects_are_equal
,
rmf
,
rmf
,
...
...
tests/nn/data_parallel/test_fsdp_shared_weights.py
View file @
8f8f8ef9
...
@@ -17,7 +17,13 @@ import torch.multiprocessing as mp
...
@@ -17,7 +17,13 @@ import torch.multiprocessing as mp
from
torch.nn
import
Linear
,
Module
from
torch.nn
import
Linear
,
Module
from
torch.optim
import
SGD
from
torch.optim
import
SGD
from
fair_dev.testing.testing
import
dist_init
,
objects_are_equal
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
from
fairscale.fair_dev.testing.testing
import
(
dist_init
,
objects_are_equal
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
,
)
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
...
...
tests/nn/data_parallel/test_fsdp_shared_weights_mevo.py
View file @
8f8f8ef9
...
@@ -17,7 +17,8 @@ from torch import nn
...
@@ -17,7 +17,8 @@ from torch import nn
import
torch.multiprocessing
as
mp
import
torch.multiprocessing
as
mp
from
torch.optim
import
SGD
from
torch.optim
import
SGD
from
fair_dev.testing.testing
import
(
from
fairscale.experimental.nn
import
MEVO
from
fairscale.fair_dev.testing.testing
import
(
dist_init
,
dist_init
,
in_circle_ci
,
in_circle_ci
,
objects_are_equal
,
objects_are_equal
,
...
@@ -25,7 +26,6 @@ from fair_dev.testing.testing import (
...
@@ -25,7 +26,6 @@ from fair_dev.testing.testing import (
teardown
,
teardown
,
temp_files_ctx
,
temp_files_ctx
,
)
)
from
fairscale.experimental.nn
import
MEVO
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
VOCAB
=
4
VOCAB
=
4
...
...
tests/nn/data_parallel/test_fsdp_state_dict.py
View file @
8f8f8ef9
...
@@ -11,7 +11,7 @@ import pytest
...
@@ -11,7 +11,7 @@ import pytest
import
torch
import
torch
from
torch
import
nn
from
torch
import
nn
from
fair_dev.testing.testing
import
dist_init
,
objects_are_equal
,
skip_if_cuda
,
teardown
,
temp_files_ctx
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
objects_are_equal
,
skip_if_cuda
,
teardown
,
temp_files_ctx
from
fairscale.internal
import
torch_version
from
fairscale.internal
import
torch_version
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
...
...
tests/nn/data_parallel/test_fsdp_uneven.py
View file @
8f8f8ef9
...
@@ -18,7 +18,7 @@ import torch.multiprocessing as mp
...
@@ -18,7 +18,7 @@ import torch.multiprocessing as mp
from
torch.nn
import
Linear
,
Sequential
from
torch.nn
import
Linear
,
Sequential
from
torch.optim
import
SGD
from
torch.optim
import
SGD
from
fair_dev.testing.testing
import
dist_init
,
skip_if_single_gpu
,
teardown
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
skip_if_single_gpu
,
teardown
from
fairscale.internal
import
torch_version
from
fairscale.internal
import
torch_version
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel.fully_sharded_data_parallel
import
TrainingState
from
fairscale.nn.data_parallel.fully_sharded_data_parallel
import
TrainingState
...
...
tests/nn/data_parallel/test_fsdp_with_checkpoint_wrapper.py
View file @
8f8f8ef9
...
@@ -13,7 +13,7 @@ from torch import nn
...
@@ -13,7 +13,7 @@ from torch import nn
import
torch.distributed
import
torch.distributed
import
torch.multiprocessing
as
mp
import
torch.multiprocessing
as
mp
from
fair_dev.testing.testing
import
dist_init
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
skip_if_single_gpu
,
teardown
,
temp_files_ctx
from
fairscale.nn.checkpoint.checkpoint_activations
import
checkpoint_wrapper
from
fairscale.nn.checkpoint.checkpoint_activations
import
checkpoint_wrapper
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
...
...
tests/nn/data_parallel/test_sharded_ddp_features.py
View file @
8f8f8ef9
...
@@ -16,7 +16,7 @@ import torch.distributed as dist
...
@@ -16,7 +16,7 @@ import torch.distributed as dist
import
torch.multiprocessing
as
mp
import
torch.multiprocessing
as
mp
from
torch.nn
import
Linear
,
Sequential
from
torch.nn
import
Linear
,
Sequential
from
fair_dev.testing.testing
import
(
from
fairscale.
fair_dev.testing.testing
import
(
GPT2
,
GPT2
,
SGDWithPausingCompute
,
SGDWithPausingCompute
,
available_devices
,
available_devices
,
...
...
Prev
1
2
3
4
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment