"src/vscode:/vscode.git/clone" did not exist on "4f438de35a88a85175c0e124b993188953bbd6e9"
Unverified Commit 8f8f8ef9 authored by Min Xu's avatar Min Xu Committed by GitHub
Browse files

[chore] move fair_dev into fairscale (#1078)


Co-authored-by: default avatarMin Xu <min.xu.public@gmail.com>
parent bfd57ff3
...@@ -19,7 +19,12 @@ import torch.multiprocessing as mp ...@@ -19,7 +19,12 @@ import torch.multiprocessing as mp
from torch.nn import Linear, Sequential from torch.nn import Linear, Sequential
from torch.nn.parallel import DistributedDataParallel as DDP from torch.nn.parallel import DistributedDataParallel as DDP
from fair_dev.testing.testing import check_same_model_params, skip_if_no_cuda, skip_if_single_gpu, temp_files_ctx from fairscale.fair_dev.testing.testing import (
check_same_model_params,
skip_if_no_cuda,
skip_if_single_gpu,
temp_files_ctx,
)
from fairscale.internal import torch_version from fairscale.internal import torch_version
from fairscale.nn.data_parallel import ShardedDataParallel from fairscale.nn.data_parallel import ShardedDataParallel
from fairscale.optim import OSS from fairscale.optim import OSS
......
...@@ -10,7 +10,7 @@ import unittest ...@@ -10,7 +10,7 @@ import unittest
import torch import torch
from fair_dev.testing.testing import objects_are_equal from fairscale.fair_dev.testing.testing import objects_are_equal
from fairscale.nn import FlattenParamsWrapper from fairscale.nn import FlattenParamsWrapper
......
...@@ -23,7 +23,7 @@ ...@@ -23,7 +23,7 @@
import torch import torch
import torch.nn.functional as F import torch.nn.functional as F
from fair_dev.testing.testing import IdentityLayer, dist_init, set_random_seed, spawn_for_all_world_sizes from fairscale.fair_dev.testing.testing import IdentityLayer, dist_init, set_random_seed, spawn_for_all_world_sizes
from fairscale.nn.model_parallel import initialize as mpu from fairscale.nn.model_parallel import initialize as mpu
from fairscale.nn.model_parallel.cross_entropy import vocab_parallel_cross_entropy from fairscale.nn.model_parallel.cross_entropy import vocab_parallel_cross_entropy
from fairscale.nn.model_parallel.mappings import scatter_to_model_parallel_region from fairscale.nn.model_parallel.mappings import scatter_to_model_parallel_region
......
...@@ -22,7 +22,7 @@ ...@@ -22,7 +22,7 @@
import torch import torch
from fair_dev.testing.testing import dist_init, spawn_for_all_world_sizes from fairscale.fair_dev.testing.testing import dist_init, spawn_for_all_world_sizes
from fairscale.nn.model_parallel import initialize as mpu from fairscale.nn.model_parallel import initialize as mpu
......
...@@ -24,7 +24,7 @@ import torch ...@@ -24,7 +24,7 @@ import torch
import torch.nn.init as init import torch.nn.init as init
from torch.nn.parameter import Parameter from torch.nn.parameter import Parameter
from fair_dev.testing.testing import dist_init, set_random_seed, spawn_for_all_world_sizes from fairscale.fair_dev.testing.testing import dist_init, set_random_seed, spawn_for_all_world_sizes
from fairscale.nn.model_parallel import initialize as mpu from fairscale.nn.model_parallel import initialize as mpu
from fairscale.nn.model_parallel import layers from fairscale.nn.model_parallel import layers
......
...@@ -21,7 +21,7 @@ ...@@ -21,7 +21,7 @@
import torch import torch
from fair_dev.testing.testing import dist_init, spawn_for_all_world_sizes from fairscale.fair_dev.testing.testing import dist_init, spawn_for_all_world_sizes
from fairscale.nn.model_parallel import initialize as mpu from fairscale.nn.model_parallel import initialize as mpu
from fairscale.nn.model_parallel import random from fairscale.nn.model_parallel import random
from fairscale.nn.model_parallel.random import get_cuda_rng_tracker, model_parallel_cuda_manual_seed from fairscale.nn.model_parallel.random import get_cuda_rng_tracker, model_parallel_cuda_manual_seed
......
...@@ -11,7 +11,7 @@ import torch ...@@ -11,7 +11,7 @@ import torch
import torch.distributed as dist import torch.distributed as dist
import torch.multiprocessing as mp import torch.multiprocessing as mp
from fair_dev.testing.testing import make_cudnn_deterministic from fairscale.fair_dev.testing.testing import make_cudnn_deterministic
from fairscale.internal import torch_version from fairscale.internal import torch_version
from fairscale.nn import MOELayer, Top2Gate from fairscale.nn import MOELayer, Top2Gate
......
...@@ -21,7 +21,7 @@ import pytest ...@@ -21,7 +21,7 @@ import pytest
import torch import torch
from torch import nn from torch import nn
from fair_dev.testing.testing import skip_if_single_gpu from fairscale.fair_dev.testing.testing import skip_if_single_gpu
from fairscale.nn.pipe import Pipe from fairscale.nn.pipe import Pipe
from fairscale.nn.pipe.skip import pop, skippable, stash from fairscale.nn.pipe.skip import pop, skippable, stash
from fairscale.nn.pipe.skip.portal import PortalBlue, PortalCopy, PortalOrange from fairscale.nn.pipe.skip.portal import PortalBlue, PortalCopy, PortalOrange
......
...@@ -22,7 +22,7 @@ import torch ...@@ -22,7 +22,7 @@ import torch
from torch import nn from torch import nn
import torch.nn.functional as F import torch.nn.functional as F
from fair_dev.testing.testing import skip_if_single_gpu from fairscale.fair_dev.testing.testing import skip_if_single_gpu
from fairscale.nn.pipe import Pipe from fairscale.nn.pipe import Pipe
......
...@@ -20,7 +20,7 @@ from torch.nn import Linear, Sequential ...@@ -20,7 +20,7 @@ from torch.nn import Linear, Sequential
from torch.nn.parallel import DistributedDataParallel as DDP from torch.nn.parallel import DistributedDataParallel as DDP
from torch.utils.checkpoint import checkpoint as torch_checkpoint from torch.utils.checkpoint import checkpoint as torch_checkpoint
from fair_dev.testing.testing import skip_if_no_cuda, skip_if_single_gpu from fairscale.fair_dev.testing.testing import skip_if_no_cuda, skip_if_single_gpu
from fairscale.nn.pipe.checkpoint import Checkpointing, Function, TensorOrTensors from fairscale.nn.pipe.checkpoint import Checkpointing, Function, TensorOrTensors
from fairscale.nn.pipe.microbatch import Batch from fairscale.nn.pipe.microbatch import Batch
......
...@@ -14,7 +14,7 @@ import numpy as np ...@@ -14,7 +14,7 @@ import numpy as np
import pytest import pytest
import torch import torch
from fair_dev.testing.testing import skip_if_single_gpu from fairscale.fair_dev.testing.testing import skip_if_single_gpu
from fairscale.nn import Pipe from fairscale.nn import Pipe
......
...@@ -22,7 +22,7 @@ import torch ...@@ -22,7 +22,7 @@ import torch
from torch import nn from torch import nn
import torch.nn.functional as F import torch.nn.functional as F
from fair_dev.testing.testing import get_worker_map, torch_spawn from fairscale.fair_dev.testing.testing import get_worker_map, torch_spawn
from fairscale.nn.pipe import AsyncPipe from fairscale.nn.pipe import AsyncPipe
......
...@@ -21,7 +21,7 @@ import pytest ...@@ -21,7 +21,7 @@ import pytest
import torch import torch
from torch import nn from torch import nn
from fair_dev.testing.testing import get_worker_map, torch_spawn from fairscale.fair_dev.testing.testing import get_worker_map, torch_spawn
from fairscale.nn.pipe import AsyncPipe from fairscale.nn.pipe import AsyncPipe
......
...@@ -26,7 +26,7 @@ import pytest ...@@ -26,7 +26,7 @@ import pytest
import torch import torch
from torch import nn from torch import nn
from fair_dev.testing.testing import get_worker_map, torch_spawn from fairscale.fair_dev.testing.testing import get_worker_map, torch_spawn
from fairscale.internal import torch_version from fairscale.internal import torch_version
from fairscale.nn.model_parallel.initialize import get_pipeline_parallel_group from fairscale.nn.model_parallel.initialize import get_pipeline_parallel_group
from fairscale.nn.pipe import AsyncPipe from fairscale.nn.pipe import AsyncPipe
......
...@@ -6,7 +6,7 @@ import torch ...@@ -6,7 +6,7 @@ import torch
from torch import nn from torch import nn
from torch.distributed import rpc from torch.distributed import rpc
from fair_dev.testing.testing import get_worker_map, torch_spawn from fairscale.fair_dev.testing.testing import get_worker_map, torch_spawn
from fairscale.internal import torch_version from fairscale.internal import torch_version
from fairscale.nn.model_parallel.initialize import get_pipeline_parallel_group from fairscale.nn.model_parallel.initialize import get_pipeline_parallel_group
from fairscale.nn.pipe import PipeRPCWrapper from fairscale.nn.pipe import PipeRPCWrapper
......
...@@ -21,7 +21,7 @@ import pytest ...@@ -21,7 +21,7 @@ import pytest
import torch import torch
from torch import nn from torch import nn
from fair_dev.testing.testing import get_worker_map, set_random_seed, torch_spawn from fairscale.fair_dev.testing.testing import get_worker_map, set_random_seed, torch_spawn
from fairscale.nn.pipe import AsyncPipe from fairscale.nn.pipe import AsyncPipe
......
...@@ -12,7 +12,7 @@ import torch ...@@ -12,7 +12,7 @@ import torch
import torch.nn as nn import torch.nn as nn
import torch.nn.functional as F import torch.nn.functional as F
from fair_dev.testing.testing import DummyProcessGroup from fairscale.fair_dev.testing.testing import DummyProcessGroup
from fairscale.nn import FullyShardedDataParallel as FSDP from fairscale.nn import FullyShardedDataParallel as FSDP
from fairscale.nn import auto_wrap, default_auto_wrap_policy, enable_wrap, wrap from fairscale.nn import auto_wrap, default_auto_wrap_policy, enable_wrap, wrap
......
...@@ -33,8 +33,8 @@ from torch.nn import Linear ...@@ -33,8 +33,8 @@ from torch.nn import Linear
from torch.nn.parallel import DistributedDataParallel as DDP from torch.nn.parallel import DistributedDataParallel as DDP
from torch.optim import SGD from torch.optim import SGD
from fair_dev.testing.golden_testing_data import adascale_test_data from fairscale.fair_dev.testing.golden_testing_data import adascale_test_data
from fair_dev.testing.testing import skip_if_single_gpu from fairscale.fair_dev.testing.testing import skip_if_single_gpu
from fairscale.nn.data_parallel import FullyShardedDataParallel as FSDP from fairscale.nn.data_parallel import FullyShardedDataParallel as FSDP
from fairscale.nn.data_parallel import ShardedDataParallel as SDP from fairscale.nn.data_parallel import ShardedDataParallel as SDP
from fairscale.optim import OSS, AdaScale from fairscale.optim import OSS, AdaScale
......
...@@ -17,8 +17,8 @@ from torch.utils.data import DataLoader ...@@ -17,8 +17,8 @@ from torch.utils.data import DataLoader
import torchvision import torchvision
import torchvision.transforms as transforms import torchvision.transforms as transforms
from fair_dev.common_paths import DATASET_CACHE_ROOT from fairscale.fair_dev.common_paths import DATASET_CACHE_ROOT
from fair_dev.testing.testing import skip_a_test_if_in_CI from fairscale.fair_dev.testing.testing import skip_a_test_if_in_CI
from fairscale.optim.layerwise_gradient_scaler import LayerwiseGradientScaler from fairscale.optim.layerwise_gradient_scaler import LayerwiseGradientScaler
......
...@@ -21,7 +21,7 @@ import torch.distributed as dist ...@@ -21,7 +21,7 @@ import torch.distributed as dist
import torch.multiprocessing as mp import torch.multiprocessing as mp
from torch.nn.parallel import DistributedDataParallel as DDP from torch.nn.parallel import DistributedDataParallel as DDP
from fair_dev.testing.testing import ( from fairscale.fair_dev.testing.testing import (
check_same_model_params, check_same_model_params,
check_same_models_across_ranks, check_same_models_across_ranks,
skip_if_no_cuda, skip_if_no_cuda,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment