Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
fairscale
Commits
8f8f8ef9
Unverified
Commit
8f8f8ef9
authored
Sep 23, 2022
by
Min Xu
Committed by
GitHub
Sep 23, 2022
Browse files
[chore] move fair_dev into fairscale (#1078)
Co-authored-by:
Min Xu
<
min.xu.public@gmail.com
>
parent
bfd57ff3
Changes
63
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
27 additions
and
22 deletions
+27
-22
tests/nn/data_parallel/test_sharded_ddp_pytorch_parity.py
tests/nn/data_parallel/test_sharded_ddp_pytorch_parity.py
+6
-1
tests/nn/misc/test_flatten_params_wrapper.py
tests/nn/misc/test_flatten_params_wrapper.py
+1
-1
tests/nn/model_parallel/test_cross_entropy.py
tests/nn/model_parallel/test_cross_entropy.py
+1
-1
tests/nn/model_parallel/test_initialize.py
tests/nn/model_parallel/test_initialize.py
+1
-1
tests/nn/model_parallel/test_layers.py
tests/nn/model_parallel/test_layers.py
+1
-1
tests/nn/model_parallel/test_random.py
tests/nn/model_parallel/test_random.py
+1
-1
tests/nn/moe/test_moe_layer.py
tests/nn/moe/test_moe_layer.py
+1
-1
tests/nn/pipe/skip/test_gpipe.py
tests/nn/pipe/skip/test_gpipe.py
+1
-1
tests/nn/pipe/test_bugs.py
tests/nn/pipe/test_bugs.py
+1
-1
tests/nn/pipe/test_checkpoint_ddp.py
tests/nn/pipe/test_checkpoint_ddp.py
+1
-1
tests/nn/pipe/test_parity.py
tests/nn/pipe/test_parity.py
+1
-1
tests/nn/pipe_process/test_bugs.py
tests/nn/pipe_process/test_bugs.py
+1
-1
tests/nn/pipe_process/test_inplace.py
tests/nn/pipe_process/test_inplace.py
+1
-1
tests/nn/pipe_process/test_pipe.py
tests/nn/pipe_process/test_pipe.py
+1
-1
tests/nn/pipe_process/test_rpc.py
tests/nn/pipe_process/test_rpc.py
+1
-1
tests/nn/pipe_process/test_transparency.py
tests/nn/pipe_process/test_transparency.py
+1
-1
tests/nn/wrap/test_wrap.py
tests/nn/wrap/test_wrap.py
+1
-1
tests/optim/test_ddp_adascale.py
tests/optim/test_ddp_adascale.py
+2
-2
tests/optim/test_layerwise_gradient_scaler.py
tests/optim/test_layerwise_gradient_scaler.py
+2
-2
tests/optim/test_oss.py
tests/optim/test_oss.py
+1
-1
No files found.
tests/nn/data_parallel/test_sharded_ddp_pytorch_parity.py
View file @
8f8f8ef9
...
...
@@ -19,7 +19,12 @@ import torch.multiprocessing as mp
from
torch.nn
import
Linear
,
Sequential
from
torch.nn.parallel
import
DistributedDataParallel
as
DDP
from
fair_dev.testing.testing
import
check_same_model_params
,
skip_if_no_cuda
,
skip_if_single_gpu
,
temp_files_ctx
from
fairscale.fair_dev.testing.testing
import
(
check_same_model_params
,
skip_if_no_cuda
,
skip_if_single_gpu
,
temp_files_ctx
,
)
from
fairscale.internal
import
torch_version
from
fairscale.nn.data_parallel
import
ShardedDataParallel
from
fairscale.optim
import
OSS
...
...
tests/nn/misc/test_flatten_params_wrapper.py
View file @
8f8f8ef9
...
...
@@ -10,7 +10,7 @@ import unittest
import
torch
from
fair_dev.testing.testing
import
objects_are_equal
from
fairscale.
fair_dev.testing.testing
import
objects_are_equal
from
fairscale.nn
import
FlattenParamsWrapper
...
...
tests/nn/model_parallel/test_cross_entropy.py
View file @
8f8f8ef9
...
...
@@ -23,7 +23,7 @@
import
torch
import
torch.nn.functional
as
F
from
fair_dev.testing.testing
import
IdentityLayer
,
dist_init
,
set_random_seed
,
spawn_for_all_world_sizes
from
fairscale.
fair_dev.testing.testing
import
IdentityLayer
,
dist_init
,
set_random_seed
,
spawn_for_all_world_sizes
from
fairscale.nn.model_parallel
import
initialize
as
mpu
from
fairscale.nn.model_parallel.cross_entropy
import
vocab_parallel_cross_entropy
from
fairscale.nn.model_parallel.mappings
import
scatter_to_model_parallel_region
...
...
tests/nn/model_parallel/test_initialize.py
View file @
8f8f8ef9
...
...
@@ -22,7 +22,7 @@
import
torch
from
fair_dev.testing.testing
import
dist_init
,
spawn_for_all_world_sizes
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
spawn_for_all_world_sizes
from
fairscale.nn.model_parallel
import
initialize
as
mpu
...
...
tests/nn/model_parallel/test_layers.py
View file @
8f8f8ef9
...
...
@@ -24,7 +24,7 @@ import torch
import
torch.nn.init
as
init
from
torch.nn.parameter
import
Parameter
from
fair_dev.testing.testing
import
dist_init
,
set_random_seed
,
spawn_for_all_world_sizes
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
set_random_seed
,
spawn_for_all_world_sizes
from
fairscale.nn.model_parallel
import
initialize
as
mpu
from
fairscale.nn.model_parallel
import
layers
...
...
tests/nn/model_parallel/test_random.py
View file @
8f8f8ef9
...
...
@@ -21,7 +21,7 @@
import
torch
from
fair_dev.testing.testing
import
dist_init
,
spawn_for_all_world_sizes
from
fairscale.
fair_dev.testing.testing
import
dist_init
,
spawn_for_all_world_sizes
from
fairscale.nn.model_parallel
import
initialize
as
mpu
from
fairscale.nn.model_parallel
import
random
from
fairscale.nn.model_parallel.random
import
get_cuda_rng_tracker
,
model_parallel_cuda_manual_seed
...
...
tests/nn/moe/test_moe_layer.py
View file @
8f8f8ef9
...
...
@@ -11,7 +11,7 @@ import torch
import
torch.distributed
as
dist
import
torch.multiprocessing
as
mp
from
fair_dev.testing.testing
import
make_cudnn_deterministic
from
fairscale.
fair_dev.testing.testing
import
make_cudnn_deterministic
from
fairscale.internal
import
torch_version
from
fairscale.nn
import
MOELayer
,
Top2Gate
...
...
tests/nn/pipe/skip/test_gpipe.py
View file @
8f8f8ef9
...
...
@@ -21,7 +21,7 @@ import pytest
import
torch
from
torch
import
nn
from
fair_dev.testing.testing
import
skip_if_single_gpu
from
fairscale.
fair_dev.testing.testing
import
skip_if_single_gpu
from
fairscale.nn.pipe
import
Pipe
from
fairscale.nn.pipe.skip
import
pop
,
skippable
,
stash
from
fairscale.nn.pipe.skip.portal
import
PortalBlue
,
PortalCopy
,
PortalOrange
...
...
tests/nn/pipe/test_bugs.py
View file @
8f8f8ef9
...
...
@@ -22,7 +22,7 @@ import torch
from
torch
import
nn
import
torch.nn.functional
as
F
from
fair_dev.testing.testing
import
skip_if_single_gpu
from
fairscale.
fair_dev.testing.testing
import
skip_if_single_gpu
from
fairscale.nn.pipe
import
Pipe
...
...
tests/nn/pipe/test_checkpoint_ddp.py
View file @
8f8f8ef9
...
...
@@ -20,7 +20,7 @@ from torch.nn import Linear, Sequential
from
torch.nn.parallel
import
DistributedDataParallel
as
DDP
from
torch.utils.checkpoint
import
checkpoint
as
torch_checkpoint
from
fair_dev.testing.testing
import
skip_if_no_cuda
,
skip_if_single_gpu
from
fairscale.
fair_dev.testing.testing
import
skip_if_no_cuda
,
skip_if_single_gpu
from
fairscale.nn.pipe.checkpoint
import
Checkpointing
,
Function
,
TensorOrTensors
from
fairscale.nn.pipe.microbatch
import
Batch
...
...
tests/nn/pipe/test_parity.py
View file @
8f8f8ef9
...
...
@@ -14,7 +14,7 @@ import numpy as np
import
pytest
import
torch
from
fair_dev.testing.testing
import
skip_if_single_gpu
from
fairscale.
fair_dev.testing.testing
import
skip_if_single_gpu
from
fairscale.nn
import
Pipe
...
...
tests/nn/pipe_process/test_bugs.py
View file @
8f8f8ef9
...
...
@@ -22,7 +22,7 @@ import torch
from
torch
import
nn
import
torch.nn.functional
as
F
from
fair_dev.testing.testing
import
get_worker_map
,
torch_spawn
from
fairscale.
fair_dev.testing.testing
import
get_worker_map
,
torch_spawn
from
fairscale.nn.pipe
import
AsyncPipe
...
...
tests/nn/pipe_process/test_inplace.py
View file @
8f8f8ef9
...
...
@@ -21,7 +21,7 @@ import pytest
import
torch
from
torch
import
nn
from
fair_dev.testing.testing
import
get_worker_map
,
torch_spawn
from
fairscale.
fair_dev.testing.testing
import
get_worker_map
,
torch_spawn
from
fairscale.nn.pipe
import
AsyncPipe
...
...
tests/nn/pipe_process/test_pipe.py
View file @
8f8f8ef9
...
...
@@ -26,7 +26,7 @@ import pytest
import
torch
from
torch
import
nn
from
fair_dev.testing.testing
import
get_worker_map
,
torch_spawn
from
fairscale.
fair_dev.testing.testing
import
get_worker_map
,
torch_spawn
from
fairscale.internal
import
torch_version
from
fairscale.nn.model_parallel.initialize
import
get_pipeline_parallel_group
from
fairscale.nn.pipe
import
AsyncPipe
...
...
tests/nn/pipe_process/test_rpc.py
View file @
8f8f8ef9
...
...
@@ -6,7 +6,7 @@ import torch
from
torch
import
nn
from
torch.distributed
import
rpc
from
fair_dev.testing.testing
import
get_worker_map
,
torch_spawn
from
fairscale.
fair_dev.testing.testing
import
get_worker_map
,
torch_spawn
from
fairscale.internal
import
torch_version
from
fairscale.nn.model_parallel.initialize
import
get_pipeline_parallel_group
from
fairscale.nn.pipe
import
PipeRPCWrapper
...
...
tests/nn/pipe_process/test_transparency.py
View file @
8f8f8ef9
...
...
@@ -21,7 +21,7 @@ import pytest
import
torch
from
torch
import
nn
from
fair_dev.testing.testing
import
get_worker_map
,
set_random_seed
,
torch_spawn
from
fairscale.
fair_dev.testing.testing
import
get_worker_map
,
set_random_seed
,
torch_spawn
from
fairscale.nn.pipe
import
AsyncPipe
...
...
tests/nn/wrap/test_wrap.py
View file @
8f8f8ef9
...
...
@@ -12,7 +12,7 @@ import torch
import
torch.nn
as
nn
import
torch.nn.functional
as
F
from
fair_dev.testing.testing
import
DummyProcessGroup
from
fairscale.
fair_dev.testing.testing
import
DummyProcessGroup
from
fairscale.nn
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn
import
auto_wrap
,
default_auto_wrap_policy
,
enable_wrap
,
wrap
...
...
tests/optim/test_ddp_adascale.py
View file @
8f8f8ef9
...
...
@@ -33,8 +33,8 @@ from torch.nn import Linear
from
torch.nn.parallel
import
DistributedDataParallel
as
DDP
from
torch.optim
import
SGD
from
fair_dev.testing.golden_testing_data
import
adascale_test_data
from
fair_dev.testing.testing
import
skip_if_single_gpu
from
fairscale.
fair_dev.testing.golden_testing_data
import
adascale_test_data
from
fairscale.
fair_dev.testing.testing
import
skip_if_single_gpu
from
fairscale.nn.data_parallel
import
FullyShardedDataParallel
as
FSDP
from
fairscale.nn.data_parallel
import
ShardedDataParallel
as
SDP
from
fairscale.optim
import
OSS
,
AdaScale
...
...
tests/optim/test_layerwise_gradient_scaler.py
View file @
8f8f8ef9
...
...
@@ -17,8 +17,8 @@ from torch.utils.data import DataLoader
import
torchvision
import
torchvision.transforms
as
transforms
from
fair_dev.common_paths
import
DATASET_CACHE_ROOT
from
fair_dev.testing.testing
import
skip_a_test_if_in_CI
from
fairscale.
fair_dev.common_paths
import
DATASET_CACHE_ROOT
from
fairscale.
fair_dev.testing.testing
import
skip_a_test_if_in_CI
from
fairscale.optim.layerwise_gradient_scaler
import
LayerwiseGradientScaler
...
...
tests/optim/test_oss.py
View file @
8f8f8ef9
...
...
@@ -21,7 +21,7 @@ import torch.distributed as dist
import
torch.multiprocessing
as
mp
from
torch.nn.parallel
import
DistributedDataParallel
as
DDP
from
fair_dev.testing.testing
import
(
from
fairscale.
fair_dev.testing.testing
import
(
check_same_model_params
,
check_same_models_across_ranks
,
skip_if_no_cuda
,
...
...
Prev
1
2
3
4
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment