"vscode:/vscode.git/clone" did not exist on "64cb699e8ae8d8c31984d9654f1de0f5e8618e4a"
Commit 71bc2f7a authored by mibaumgartner's avatar mibaumgartner
Browse files

rename models -> arch

parent 4560ce77
from nndet.arch.blocks.basic import AbstractBlock, StackedConvBlock, \
StackedResidualBlock, StackedConvBlock2
from nndet.arch.blocks.res import ResBasic, ResBottleneck
...@@ -20,8 +20,8 @@ import torch.nn as nn ...@@ -20,8 +20,8 @@ import torch.nn as nn
from abc import abstractmethod from abc import abstractmethod
from typing import Sequence, Callable, Union, Tuple from typing import Sequence, Callable, Union, Tuple
from nndet.models.conv import NdParam from nndet.arch.conv import NdParam
from nndet.models.blocks.res import ResBasic from nndet.arch.blocks.res import ResBasic
class AbstractBlock(nn.Module): class AbstractBlock(nn.Module):
......
...@@ -21,8 +21,8 @@ from typing import Sequence, Callable, Optional ...@@ -21,8 +21,8 @@ from typing import Sequence, Callable, Optional
from functools import reduce from functools import reduce
from loguru import logger from loguru import logger
from nndet.models.conv import nd_pool from nndet.arch.conv import nd_pool
from nndet.models.conv import NdParam from nndet.arch.conv import NdParam
class ResBasic(nn.Module): class ResBasic(nn.Module):
......
...@@ -18,7 +18,7 @@ import torch ...@@ -18,7 +18,7 @@ import torch
import torch.nn as nn import torch.nn as nn
from nndet.models.conv import nd_pool, nd_conv from nndet.arch.conv import nd_pool, nd_conv
class SELayer(nn.Module): class SELayer(nn.Module):
......
...@@ -18,8 +18,8 @@ import torch ...@@ -18,8 +18,8 @@ import torch
import torch.nn as nn import torch.nn as nn
from typing import Union, Callable, Any, Optional, Tuple, Sequence, Type from typing import Union, Callable, Any, Optional, Tuple, Sequence, Type
from nndet.models.initializer import InitWeights_He from nndet.arch.initializer import InitWeights_He
from nndet.models.layers.norm import GroupNorm from nndet.arch.layers.norm import GroupNorm
NdParam = Union[int, Tuple[int, int], Tuple[int, int, int]] NdParam = Union[int, Tuple[int, int], Tuple[int, int, int]]
...@@ -415,7 +415,7 @@ def nd_norm(norm_type: str, dim: int, *args, **kwargs) -> torch.nn.Module: ...@@ -415,7 +415,7 @@ def nd_norm(norm_type: str, dim: int, *args, **kwargs) -> torch.nn.Module:
* :class:`torch.nn.InstanceNorm2d` * :class:`torch.nn.InstanceNorm2d`
* :class:`torch.nn.InstanceNorm3d` * :class:`torch.nn.InstanceNorm3d`
* :class:`torch.nn.LocalResponseNorm` * :class:`torch.nn.LocalResponseNorm`
* :class:`nndet.models.layers.norm.GroupNorm` * :class:`nndet.arch.layers.norm.GroupNorm`
""" """
if dim is None: if dim is None:
dim_str = "" dim_str = ""
......
from nndet.arch.decoder.fpn import FPN, UFPN, FPN2
...@@ -20,7 +20,7 @@ from typing import Sequence, List, Tuple, Union, Callable, Optional, TypeVar ...@@ -20,7 +20,7 @@ from typing import Sequence, List, Tuple, Union, Callable, Optional, TypeVar
from loguru import logger from loguru import logger
from nndet.models.conv import conv_kwargs_helper from nndet.arch.conv import conv_kwargs_helper
from nndet.utils import to_dtype from nndet.utils import to_dtype
......
from nndet.arch.encoder.abstract import AbstractEncoder
from nndet.arch.encoder.modular import Encoder
...@@ -18,8 +18,8 @@ import torch ...@@ -18,8 +18,8 @@ import torch
import torch.nn as nn import torch.nn as nn
from typing import Callable, Tuple, Sequence, Union, List, Dict, Optional from typing import Callable, Tuple, Sequence, Union, List, Dict, Optional
from nndet.models.encoder.abstract import AbstractEncoder from nndet.arch.encoder.abstract import AbstractEncoder
from nndet.models.blocks.basic import AbstractBlock from nndet.arch.blocks.basic import AbstractBlock
__all__ = ["Encoder"] __all__ = ["Encoder"]
......
from nndet.arch.heads.classifier import ClassifierType, Classifier
from nndet.arch.heads.comb import HeadType, AbstractHead
from nndet.arch.heads.regressor import RegressorType, Regressor
from nndet.arch.heads.segmenter import SegmenterType, Segmenter
...@@ -23,8 +23,8 @@ from abc import abstractmethod ...@@ -23,8 +23,8 @@ from abc import abstractmethod
from nndet.detection.boxes import BoxCoderND from nndet.detection.boxes import BoxCoderND
from nndet.detection.boxes.sampler import AbstractSampler from nndet.detection.boxes.sampler import AbstractSampler
from nndet.models.heads.classifier import Classifier from nndet.arch.heads.classifier import Classifier
from nndet.models.heads.regressor import Regressor from nndet.arch.heads.regressor import Regressor
class AbstractHead(nn.Module): class AbstractHead(nn.Module):
......
...@@ -23,7 +23,7 @@ from abc import abstractmethod ...@@ -23,7 +23,7 @@ from abc import abstractmethod
from loguru import logger from loguru import logger
from nndet.detection.boxes import box_iou from nndet.detection.boxes import box_iou
from nndet.models.layers.scale import Scale from nndet.arch.layers.scale import Scale
from torch import Tensor from torch import Tensor
from nndet.losses import SmoothL1Loss, GIoULoss from nndet.losses import SmoothL1Loss, GIoULoss
......
...@@ -20,10 +20,10 @@ import torch.nn as nn ...@@ -20,10 +20,10 @@ import torch.nn as nn
from torch import Tensor from torch import Tensor
from typing import Dict, List, Union, Sequence, Optional, Tuple, TypeVar from typing import Dict, List, Union, Sequence, Optional, Tuple, TypeVar
from nndet.models.conv import compute_padding_for_kernel, conv_kwargs_helper from nndet.arch.conv import compute_padding_for_kernel, conv_kwargs_helper
from nndet.models.heads.comb import AbstractHead from nndet.arch.heads.comb import AbstractHead
from nndet.arch.layers.interpolation import InterpolateToShapes
from nndet.losses.segmentation import SoftDiceLoss, TopKLoss from nndet.losses.segmentation import SoftDiceLoss, TopKLoss
from nndet.models.layers.interpolation import InterpolateToShapes
class Segmenter(AbstractHead): class Segmenter(AbstractHead):
......
from nndet.models.layers.interpolation import ( from nndet.arch.layers.interpolation import (
Interpolate, Interpolate,
InterpolateToShapes, InterpolateToShapes,
InterpolateToShape, InterpolateToShape,
MaxPoolToShapes, MaxPoolToShapes,
) )
from nndet.models.layers.norm import GroupNorm from nndet.arch.layers.norm import GroupNorm
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment