Unverified Commit 5f0edb97 authored by Philip Meier's avatar Philip Meier Committed by GitHub
Browse files

Add ufmt (usort + black) as code formatter (#4384)



* add ufmt as code formatter

* cleanup

* quote ufmt requirement

* split imports into more groups

* regenerate circleci config

* fix CI

* clarify local testing utils section

* use ufmt pre-commit hook

* split relative imports into local category

* Revert "split relative imports into local category"

This reverts commit f2e224cde2008c56c9347c1f69746d39065cdd51.

* pin black and usort dependencies

* fix local test utils detection

* fix ufmt rev

* add reference utils to local category

* fix usort config

* remove custom categories sorting

* Run pre-commit without fixing flake8

* got a double import in merge
Co-authored-by: default avatarNicolas Hug <nicolashug@fb.com>
parent e45489b1
import bisect import bisect
from collections import defaultdict
import copy import copy
from itertools import repeat, chain
import math import math
import numpy as np from collections import defaultdict
from itertools import repeat, chain
import numpy as np
import torch import torch
import torch.utils.data import torch.utils.data
from torch.utils.data.sampler import BatchSampler, Sampler
from torch.utils.model_zoo import tqdm
import torchvision import torchvision
from PIL import Image from PIL import Image
from torch.utils.data.sampler import BatchSampler, Sampler
from torch.utils.model_zoo import tqdm
def _repeat_to_at_least(iterable, n): def _repeat_to_at_least(iterable, n):
...@@ -34,11 +33,11 @@ class GroupedBatchSampler(BatchSampler): ...@@ -34,11 +33,11 @@ class GroupedBatchSampler(BatchSampler):
0, i.e. they must be in the range [0, num_groups). 0, i.e. they must be in the range [0, num_groups).
batch_size (int): Size of mini-batch. batch_size (int): Size of mini-batch.
""" """
def __init__(self, sampler, group_ids, batch_size): def __init__(self, sampler, group_ids, batch_size):
if not isinstance(sampler, Sampler): if not isinstance(sampler, Sampler):
raise ValueError( raise ValueError(
"sampler should be an instance of " "sampler should be an instance of " "torch.utils.data.Sampler, but got sampler={}".format(sampler)
"torch.utils.data.Sampler, but got sampler={}".format(sampler)
) )
self.sampler = sampler self.sampler = sampler
self.group_ids = group_ids self.group_ids = group_ids
...@@ -68,8 +67,7 @@ class GroupedBatchSampler(BatchSampler): ...@@ -68,8 +67,7 @@ class GroupedBatchSampler(BatchSampler):
if num_remaining > 0: if num_remaining > 0:
# for the remaining batches, take first the buffers with largest number # for the remaining batches, take first the buffers with largest number
# of elements # of elements
for group_id, _ in sorted(buffer_per_group.items(), for group_id, _ in sorted(buffer_per_group.items(), key=lambda x: len(x[1]), reverse=True):
key=lambda x: len(x[1]), reverse=True):
remaining = self.batch_size - len(buffer_per_group[group_id]) remaining = self.batch_size - len(buffer_per_group[group_id])
samples_from_group_id = _repeat_to_at_least(samples_per_group[group_id], remaining) samples_from_group_id = _repeat_to_at_least(samples_per_group[group_id], remaining)
buffer_per_group[group_id].extend(samples_from_group_id[:remaining]) buffer_per_group[group_id].extend(samples_from_group_id[:remaining])
...@@ -85,10 +83,12 @@ class GroupedBatchSampler(BatchSampler): ...@@ -85,10 +83,12 @@ class GroupedBatchSampler(BatchSampler):
def _compute_aspect_ratios_slow(dataset, indices=None): def _compute_aspect_ratios_slow(dataset, indices=None):
print("Your dataset doesn't support the fast path for " print(
"computing the aspect ratios, so will iterate over " "Your dataset doesn't support the fast path for "
"the full dataset and load every image instead. " "computing the aspect ratios, so will iterate over "
"This might take some time...") "the full dataset and load every image instead. "
"This might take some time..."
)
if indices is None: if indices is None:
indices = range(len(dataset)) indices = range(len(dataset))
...@@ -104,9 +104,12 @@ def _compute_aspect_ratios_slow(dataset, indices=None): ...@@ -104,9 +104,12 @@ def _compute_aspect_ratios_slow(dataset, indices=None):
sampler = SubsetSampler(indices) sampler = SubsetSampler(indices)
data_loader = torch.utils.data.DataLoader( data_loader = torch.utils.data.DataLoader(
dataset, batch_size=1, sampler=sampler, dataset,
batch_size=1,
sampler=sampler,
num_workers=14, # you might want to increase it for faster processing num_workers=14, # you might want to increase it for faster processing
collate_fn=lambda x: x[0]) collate_fn=lambda x: x[0],
)
aspect_ratios = [] aspect_ratios = []
with tqdm(total=len(dataset)) as pbar: with tqdm(total=len(dataset)) as pbar:
for _i, (img, _) in enumerate(data_loader): for _i, (img, _) in enumerate(data_loader):
......
import torch import torch
import transforms as T import transforms as T
class DetectionPresetTrain: class DetectionPresetTrain:
def __init__(self, data_augmentation, hflip_prob=0.5, mean=(123., 117., 104.)): def __init__(self, data_augmentation, hflip_prob=0.5, mean=(123.0, 117.0, 104.0)):
if data_augmentation == 'hflip': if data_augmentation == "hflip":
self.transforms = T.Compose([ self.transforms = T.Compose(
T.RandomHorizontalFlip(p=hflip_prob), [
T.PILToTensor(), T.RandomHorizontalFlip(p=hflip_prob),
T.ConvertImageDtype(torch.float), T.PILToTensor(),
]) T.ConvertImageDtype(torch.float),
elif data_augmentation == 'ssd': ]
self.transforms = T.Compose([ )
T.RandomPhotometricDistort(), elif data_augmentation == "ssd":
T.RandomZoomOut(fill=list(mean)), self.transforms = T.Compose(
T.RandomIoUCrop(), [
T.RandomHorizontalFlip(p=hflip_prob), T.RandomPhotometricDistort(),
T.PILToTensor(), T.RandomZoomOut(fill=list(mean)),
T.ConvertImageDtype(torch.float), T.RandomIoUCrop(),
]) T.RandomHorizontalFlip(p=hflip_prob),
elif data_augmentation == 'ssdlite': T.PILToTensor(),
self.transforms = T.Compose([ T.ConvertImageDtype(torch.float),
T.RandomIoUCrop(), ]
T.RandomHorizontalFlip(p=hflip_prob), )
T.PILToTensor(), elif data_augmentation == "ssdlite":
T.ConvertImageDtype(torch.float), self.transforms = T.Compose(
]) [
T.RandomIoUCrop(),
T.RandomHorizontalFlip(p=hflip_prob),
T.PILToTensor(),
T.ConvertImageDtype(torch.float),
]
)
else: else:
raise ValueError(f'Unknown data augmentation policy "{data_augmentation}"') raise ValueError(f'Unknown data augmentation policy "{data_augmentation}"')
......
...@@ -21,26 +21,20 @@ import datetime ...@@ -21,26 +21,20 @@ import datetime
import os import os
import time import time
import presets
import torch import torch
import torch.utils.data import torch.utils.data
import torchvision import torchvision
import torchvision.models.detection import torchvision.models.detection
import torchvision.models.detection.mask_rcnn import torchvision.models.detection.mask_rcnn
import utils
from coco_utils import get_coco, get_coco_kp from coco_utils import get_coco, get_coco_kp
from group_by_aspect_ratio import GroupedBatchSampler, create_aspect_ratio_groups
from engine import train_one_epoch, evaluate from engine import train_one_epoch, evaluate
from group_by_aspect_ratio import GroupedBatchSampler, create_aspect_ratio_groups
import presets
import utils
def get_dataset(name, image_set, transform, data_path): def get_dataset(name, image_set, transform, data_path):
paths = { paths = {"coco": (data_path, get_coco, 91), "coco_kp": (data_path, get_coco_kp, 2)}
"coco": (data_path, get_coco, 91),
"coco_kp": (data_path, get_coco_kp, 2)
}
p, ds_fn, num_classes = paths[name] p, ds_fn, num_classes = paths[name]
ds = ds_fn(p, image_set=image_set, transforms=transform) ds = ds_fn(p, image_set=image_set, transforms=transform)
...@@ -53,42 +47,60 @@ def get_transform(train, data_augmentation): ...@@ -53,42 +47,60 @@ def get_transform(train, data_augmentation):
def get_args_parser(add_help=True): def get_args_parser(add_help=True):
import argparse import argparse
parser = argparse.ArgumentParser(description='PyTorch Detection Training', add_help=add_help)
parser = argparse.ArgumentParser(description="PyTorch Detection Training", add_help=add_help)
parser.add_argument('--data-path', default='/datasets01/COCO/022719/', help='dataset')
parser.add_argument('--dataset', default='coco', help='dataset') parser.add_argument("--data-path", default="/datasets01/COCO/022719/", help="dataset")
parser.add_argument('--model', default='maskrcnn_resnet50_fpn', help='model') parser.add_argument("--dataset", default="coco", help="dataset")
parser.add_argument('--device', default='cuda', help='device') parser.add_argument("--model", default="maskrcnn_resnet50_fpn", help="model")
parser.add_argument('-b', '--batch-size', default=2, type=int, parser.add_argument("--device", default="cuda", help="device")
help='images per gpu, the total batch size is $NGPU x batch_size') parser.add_argument(
parser.add_argument('--epochs', default=26, type=int, metavar='N', "-b", "--batch-size", default=2, type=int, help="images per gpu, the total batch size is $NGPU x batch_size"
help='number of total epochs to run') )
parser.add_argument('-j', '--workers', default=4, type=int, metavar='N', parser.add_argument("--epochs", default=26, type=int, metavar="N", help="number of total epochs to run")
help='number of data loading workers (default: 4)') parser.add_argument(
parser.add_argument('--lr', default=0.02, type=float, "-j", "--workers", default=4, type=int, metavar="N", help="number of data loading workers (default: 4)"
help='initial learning rate, 0.02 is the default value for training ' )
'on 8 gpus and 2 images_per_gpu') parser.add_argument(
parser.add_argument('--momentum', default=0.9, type=float, metavar='M', "--lr",
help='momentum') default=0.02,
parser.add_argument('--wd', '--weight-decay', default=1e-4, type=float, type=float,
metavar='W', help='weight decay (default: 1e-4)', help="initial learning rate, 0.02 is the default value for training " "on 8 gpus and 2 images_per_gpu",
dest='weight_decay') )
parser.add_argument('--lr-scheduler', default="multisteplr", help='the lr scheduler (default: multisteplr)') parser.add_argument("--momentum", default=0.9, type=float, metavar="M", help="momentum")
parser.add_argument('--lr-step-size', default=8, type=int, parser.add_argument(
help='decrease lr every step-size epochs (multisteplr scheduler only)') "--wd",
parser.add_argument('--lr-steps', default=[16, 22], nargs='+', type=int, "--weight-decay",
help='decrease lr every step-size epochs (multisteplr scheduler only)') default=1e-4,
parser.add_argument('--lr-gamma', default=0.1, type=float, type=float,
help='decrease lr by a factor of lr-gamma (multisteplr scheduler only)') metavar="W",
parser.add_argument('--print-freq', default=20, type=int, help='print frequency') help="weight decay (default: 1e-4)",
parser.add_argument('--output-dir', default='.', help='path where to save') dest="weight_decay",
parser.add_argument('--resume', default='', help='resume from checkpoint') )
parser.add_argument('--start_epoch', default=0, type=int, help='start epoch') parser.add_argument("--lr-scheduler", default="multisteplr", help="the lr scheduler (default: multisteplr)")
parser.add_argument('--aspect-ratio-group-factor', default=3, type=int) parser.add_argument(
parser.add_argument('--rpn-score-thresh', default=None, type=float, help='rpn score threshold for faster-rcnn') "--lr-step-size", default=8, type=int, help="decrease lr every step-size epochs (multisteplr scheduler only)"
parser.add_argument('--trainable-backbone-layers', default=None, type=int, )
help='number of trainable layers of backbone') parser.add_argument(
parser.add_argument('--data-augmentation', default="hflip", help='data augmentation policy (default: hflip)') "--lr-steps",
default=[16, 22],
nargs="+",
type=int,
help="decrease lr every step-size epochs (multisteplr scheduler only)",
)
parser.add_argument(
"--lr-gamma", default=0.1, type=float, help="decrease lr by a factor of lr-gamma (multisteplr scheduler only)"
)
parser.add_argument("--print-freq", default=20, type=int, help="print frequency")
parser.add_argument("--output-dir", default=".", help="path where to save")
parser.add_argument("--resume", default="", help="resume from checkpoint")
parser.add_argument("--start_epoch", default=0, type=int, help="start epoch")
parser.add_argument("--aspect-ratio-group-factor", default=3, type=int)
parser.add_argument("--rpn-score-thresh", default=None, type=float, help="rpn score threshold for faster-rcnn")
parser.add_argument(
"--trainable-backbone-layers", default=None, type=int, help="number of trainable layers of backbone"
)
parser.add_argument("--data-augmentation", default="hflip", help="data augmentation policy (default: hflip)")
parser.add_argument( parser.add_argument(
"--sync-bn", "--sync-bn",
dest="sync_bn", dest="sync_bn",
...@@ -109,9 +121,8 @@ def get_args_parser(add_help=True): ...@@ -109,9 +121,8 @@ def get_args_parser(add_help=True):
) )
# distributed training parameters # distributed training parameters
parser.add_argument('--world-size', default=1, type=int, parser.add_argument("--world-size", default=1, type=int, help="number of distributed processes")
help='number of distributed processes') parser.add_argument("--dist-url", default="env://", help="url used to set up distributed training")
parser.add_argument('--dist-url', default='env://', help='url used to set up distributed training')
return parser return parser
...@@ -128,8 +139,9 @@ def main(args): ...@@ -128,8 +139,9 @@ def main(args):
# Data loading code # Data loading code
print("Loading data") print("Loading data")
dataset, num_classes = get_dataset(args.dataset, "train", get_transform(True, args.data_augmentation), dataset, num_classes = get_dataset(
args.data_path) args.dataset, "train", get_transform(True, args.data_augmentation), args.data_path
)
dataset_test, _ = get_dataset(args.dataset, "val", get_transform(False, args.data_augmentation), args.data_path) dataset_test, _ = get_dataset(args.dataset, "val", get_transform(False, args.data_augmentation), args.data_path)
print("Creating data loaders") print("Creating data loaders")
...@@ -144,27 +156,24 @@ def main(args): ...@@ -144,27 +156,24 @@ def main(args):
group_ids = create_aspect_ratio_groups(dataset, k=args.aspect_ratio_group_factor) group_ids = create_aspect_ratio_groups(dataset, k=args.aspect_ratio_group_factor)
train_batch_sampler = GroupedBatchSampler(train_sampler, group_ids, args.batch_size) train_batch_sampler = GroupedBatchSampler(train_sampler, group_ids, args.batch_size)
else: else:
train_batch_sampler = torch.utils.data.BatchSampler( train_batch_sampler = torch.utils.data.BatchSampler(train_sampler, args.batch_size, drop_last=True)
train_sampler, args.batch_size, drop_last=True)
data_loader = torch.utils.data.DataLoader( data_loader = torch.utils.data.DataLoader(
dataset, batch_sampler=train_batch_sampler, num_workers=args.workers, dataset, batch_sampler=train_batch_sampler, num_workers=args.workers, collate_fn=utils.collate_fn
collate_fn=utils.collate_fn) )
data_loader_test = torch.utils.data.DataLoader( data_loader_test = torch.utils.data.DataLoader(
dataset_test, batch_size=1, dataset_test, batch_size=1, sampler=test_sampler, num_workers=args.workers, collate_fn=utils.collate_fn
sampler=test_sampler, num_workers=args.workers, )
collate_fn=utils.collate_fn)
print("Creating model") print("Creating model")
kwargs = { kwargs = {"trainable_backbone_layers": args.trainable_backbone_layers}
"trainable_backbone_layers": args.trainable_backbone_layers
}
if "rcnn" in args.model: if "rcnn" in args.model:
if args.rpn_score_thresh is not None: if args.rpn_score_thresh is not None:
kwargs["rpn_score_thresh"] = args.rpn_score_thresh kwargs["rpn_score_thresh"] = args.rpn_score_thresh
model = torchvision.models.detection.__dict__[args.model](num_classes=num_classes, pretrained=args.pretrained, model = torchvision.models.detection.__dict__[args.model](
**kwargs) num_classes=num_classes, pretrained=args.pretrained, **kwargs
)
model.to(device) model.to(device)
if args.distributed and args.sync_bn: if args.distributed and args.sync_bn:
model = torch.nn.SyncBatchNorm.convert_sync_batchnorm(model) model = torch.nn.SyncBatchNorm.convert_sync_batchnorm(model)
...@@ -175,24 +184,25 @@ def main(args): ...@@ -175,24 +184,25 @@ def main(args):
model_without_ddp = model.module model_without_ddp = model.module
params = [p for p in model.parameters() if p.requires_grad] params = [p for p in model.parameters() if p.requires_grad]
optimizer = torch.optim.SGD( optimizer = torch.optim.SGD(params, lr=args.lr, momentum=args.momentum, weight_decay=args.weight_decay)
params, lr=args.lr, momentum=args.momentum, weight_decay=args.weight_decay)
args.lr_scheduler = args.lr_scheduler.lower() args.lr_scheduler = args.lr_scheduler.lower()
if args.lr_scheduler == 'multisteplr': if args.lr_scheduler == "multisteplr":
lr_scheduler = torch.optim.lr_scheduler.MultiStepLR(optimizer, milestones=args.lr_steps, gamma=args.lr_gamma) lr_scheduler = torch.optim.lr_scheduler.MultiStepLR(optimizer, milestones=args.lr_steps, gamma=args.lr_gamma)
elif args.lr_scheduler == 'cosineannealinglr': elif args.lr_scheduler == "cosineannealinglr":
lr_scheduler = torch.optim.lr_scheduler.CosineAnnealingLR(optimizer, T_max=args.epochs) lr_scheduler = torch.optim.lr_scheduler.CosineAnnealingLR(optimizer, T_max=args.epochs)
else: else:
raise RuntimeError("Invalid lr scheduler '{}'. Only MultiStepLR and CosineAnnealingLR " raise RuntimeError(
"are supported.".format(args.lr_scheduler)) "Invalid lr scheduler '{}'. Only MultiStepLR and CosineAnnealingLR "
"are supported.".format(args.lr_scheduler)
)
if args.resume: if args.resume:
checkpoint = torch.load(args.resume, map_location='cpu') checkpoint = torch.load(args.resume, map_location="cpu")
model_without_ddp.load_state_dict(checkpoint['model']) model_without_ddp.load_state_dict(checkpoint["model"])
optimizer.load_state_dict(checkpoint['optimizer']) optimizer.load_state_dict(checkpoint["optimizer"])
lr_scheduler.load_state_dict(checkpoint['lr_scheduler']) lr_scheduler.load_state_dict(checkpoint["lr_scheduler"])
args.start_epoch = checkpoint['epoch'] + 1 args.start_epoch = checkpoint["epoch"] + 1
if args.test_only: if args.test_only:
evaluate(model, data_loader_test, device=device) evaluate(model, data_loader_test, device=device)
...@@ -207,25 +217,21 @@ def main(args): ...@@ -207,25 +217,21 @@ def main(args):
lr_scheduler.step() lr_scheduler.step()
if args.output_dir: if args.output_dir:
checkpoint = { checkpoint = {
'model': model_without_ddp.state_dict(), "model": model_without_ddp.state_dict(),
'optimizer': optimizer.state_dict(), "optimizer": optimizer.state_dict(),
'lr_scheduler': lr_scheduler.state_dict(), "lr_scheduler": lr_scheduler.state_dict(),
'args': args, "args": args,
'epoch': epoch "epoch": epoch,
} }
utils.save_on_master( utils.save_on_master(checkpoint, os.path.join(args.output_dir, "model_{}.pth".format(epoch)))
checkpoint, utils.save_on_master(checkpoint, os.path.join(args.output_dir, "checkpoint.pth"))
os.path.join(args.output_dir, 'model_{}.pth'.format(epoch)))
utils.save_on_master(
checkpoint,
os.path.join(args.output_dir, 'checkpoint.pth'))
# evaluate after every epoch # evaluate after every epoch
evaluate(model, data_loader_test, device=device) evaluate(model, data_loader_test, device=device)
total_time = time.time() - start_time total_time = time.time() - start_time
total_time_str = str(datetime.timedelta(seconds=int(total_time))) total_time_str = str(datetime.timedelta(seconds=int(total_time)))
print('Training time {}'.format(total_time_str)) print("Training time {}".format(total_time_str))
if __name__ == "__main__": if __name__ == "__main__":
......
...@@ -28,8 +28,9 @@ class Compose(object): ...@@ -28,8 +28,9 @@ class Compose(object):
class RandomHorizontalFlip(T.RandomHorizontalFlip): class RandomHorizontalFlip(T.RandomHorizontalFlip):
def forward(self, image: Tensor, def forward(
target: Optional[Dict[str, Tensor]] = None) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]: self, image: Tensor, target: Optional[Dict[str, Tensor]] = None
) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]:
if torch.rand(1) < self.p: if torch.rand(1) < self.p:
image = F.hflip(image) image = F.hflip(image)
if target is not None: if target is not None:
...@@ -45,16 +46,18 @@ class RandomHorizontalFlip(T.RandomHorizontalFlip): ...@@ -45,16 +46,18 @@ class RandomHorizontalFlip(T.RandomHorizontalFlip):
class ToTensor(nn.Module): class ToTensor(nn.Module):
def forward(self, image: Tensor, def forward(
target: Optional[Dict[str, Tensor]] = None) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]: self, image: Tensor, target: Optional[Dict[str, Tensor]] = None
) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]:
image = F.pil_to_tensor(image) image = F.pil_to_tensor(image)
image = F.convert_image_dtype(image) image = F.convert_image_dtype(image)
return image, target return image, target
class PILToTensor(nn.Module): class PILToTensor(nn.Module):
def forward(self, image: Tensor, def forward(
target: Optional[Dict[str, Tensor]] = None) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]: self, image: Tensor, target: Optional[Dict[str, Tensor]] = None
) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]:
image = F.pil_to_tensor(image) image = F.pil_to_tensor(image)
return image, target return image, target
...@@ -64,15 +67,23 @@ class ConvertImageDtype(nn.Module): ...@@ -64,15 +67,23 @@ class ConvertImageDtype(nn.Module):
super().__init__() super().__init__()
self.dtype = dtype self.dtype = dtype
def forward(self, image: Tensor, def forward(
target: Optional[Dict[str, Tensor]] = None) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]: self, image: Tensor, target: Optional[Dict[str, Tensor]] = None
) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]:
image = F.convert_image_dtype(image, self.dtype) image = F.convert_image_dtype(image, self.dtype)
return image, target return image, target
class RandomIoUCrop(nn.Module): class RandomIoUCrop(nn.Module):
def __init__(self, min_scale: float = 0.3, max_scale: float = 1.0, min_aspect_ratio: float = 0.5, def __init__(
max_aspect_ratio: float = 2.0, sampler_options: Optional[List[float]] = None, trials: int = 40): self,
min_scale: float = 0.3,
max_scale: float = 1.0,
min_aspect_ratio: float = 0.5,
max_aspect_ratio: float = 2.0,
sampler_options: Optional[List[float]] = None,
trials: int = 40,
):
super().__init__() super().__init__()
# Configuration similar to https://github.com/weiliu89/caffe/blob/ssd/examples/ssd/ssd_coco.py#L89-L174 # Configuration similar to https://github.com/weiliu89/caffe/blob/ssd/examples/ssd/ssd_coco.py#L89-L174
self.min_scale = min_scale self.min_scale = min_scale
...@@ -84,14 +95,15 @@ class RandomIoUCrop(nn.Module): ...@@ -84,14 +95,15 @@ class RandomIoUCrop(nn.Module):
self.options = sampler_options self.options = sampler_options
self.trials = trials self.trials = trials
def forward(self, image: Tensor, def forward(
target: Optional[Dict[str, Tensor]] = None) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]: self, image: Tensor, target: Optional[Dict[str, Tensor]] = None
) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]:
if target is None: if target is None:
raise ValueError("The targets can't be None for this transform.") raise ValueError("The targets can't be None for this transform.")
if isinstance(image, torch.Tensor): if isinstance(image, torch.Tensor):
if image.ndimension() not in {2, 3}: if image.ndimension() not in {2, 3}:
raise ValueError('image should be 2/3 dimensional. Got {} dimensions.'.format(image.ndimension())) raise ValueError("image should be 2/3 dimensional. Got {} dimensions.".format(image.ndimension()))
elif image.ndimension() == 2: elif image.ndimension() == 2:
image = image.unsqueeze(0) image = image.unsqueeze(0)
...@@ -131,8 +143,9 @@ class RandomIoUCrop(nn.Module): ...@@ -131,8 +143,9 @@ class RandomIoUCrop(nn.Module):
# check at least 1 box with jaccard limitations # check at least 1 box with jaccard limitations
boxes = target["boxes"][is_within_crop_area] boxes = target["boxes"][is_within_crop_area]
ious = torchvision.ops.boxes.box_iou(boxes, torch.tensor([[left, top, right, bottom]], ious = torchvision.ops.boxes.box_iou(
dtype=boxes.dtype, device=boxes.device)) boxes, torch.tensor([[left, top, right, bottom]], dtype=boxes.dtype, device=boxes.device)
)
if ious.max() < min_jaccard_overlap: if ious.max() < min_jaccard_overlap:
continue continue
...@@ -149,13 +162,15 @@ class RandomIoUCrop(nn.Module): ...@@ -149,13 +162,15 @@ class RandomIoUCrop(nn.Module):
class RandomZoomOut(nn.Module): class RandomZoomOut(nn.Module):
def __init__(self, fill: Optional[List[float]] = None, side_range: Tuple[float, float] = (1., 4.), p: float = 0.5): def __init__(
self, fill: Optional[List[float]] = None, side_range: Tuple[float, float] = (1.0, 4.0), p: float = 0.5
):
super().__init__() super().__init__()
if fill is None: if fill is None:
fill = [0., 0., 0.] fill = [0.0, 0.0, 0.0]
self.fill = fill self.fill = fill
self.side_range = side_range self.side_range = side_range
if side_range[0] < 1. or side_range[0] > side_range[1]: if side_range[0] < 1.0 or side_range[0] > side_range[1]:
raise ValueError("Invalid canvas side range provided {}.".format(side_range)) raise ValueError("Invalid canvas side range provided {}.".format(side_range))
self.p = p self.p = p
...@@ -165,11 +180,12 @@ class RandomZoomOut(nn.Module): ...@@ -165,11 +180,12 @@ class RandomZoomOut(nn.Module):
# We fake the type to make it work on JIT # We fake the type to make it work on JIT
return tuple(int(x) for x in self.fill) if is_pil else 0 return tuple(int(x) for x in self.fill) if is_pil else 0
def forward(self, image: Tensor, def forward(
target: Optional[Dict[str, Tensor]] = None) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]: self, image: Tensor, target: Optional[Dict[str, Tensor]] = None
) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]:
if isinstance(image, torch.Tensor): if isinstance(image, torch.Tensor):
if image.ndimension() not in {2, 3}: if image.ndimension() not in {2, 3}:
raise ValueError('image should be 2/3 dimensional. Got {} dimensions.'.format(image.ndimension())) raise ValueError("image should be 2/3 dimensional. Got {} dimensions.".format(image.ndimension()))
elif image.ndimension() == 2: elif image.ndimension() == 2:
image = image.unsqueeze(0) image = image.unsqueeze(0)
...@@ -196,8 +212,9 @@ class RandomZoomOut(nn.Module): ...@@ -196,8 +212,9 @@ class RandomZoomOut(nn.Module):
image = F.pad(image, [left, top, right, bottom], fill=fill) image = F.pad(image, [left, top, right, bottom], fill=fill)
if isinstance(image, torch.Tensor): if isinstance(image, torch.Tensor):
v = torch.tensor(self.fill, device=image.device, dtype=image.dtype).view(-1, 1, 1) v = torch.tensor(self.fill, device=image.device, dtype=image.dtype).view(-1, 1, 1)
image[..., :top, :] = image[..., :, :left] = image[..., (top + orig_h):, :] = \ image[..., :top, :] = image[..., :, :left] = image[..., (top + orig_h) :, :] = image[
image[..., :, (left + orig_w):] = v ..., :, (left + orig_w) :
] = v
if target is not None: if target is not None:
target["boxes"][:, 0::2] += left target["boxes"][:, 0::2] += left
...@@ -207,8 +224,14 @@ class RandomZoomOut(nn.Module): ...@@ -207,8 +224,14 @@ class RandomZoomOut(nn.Module):
class RandomPhotometricDistort(nn.Module): class RandomPhotometricDistort(nn.Module):
def __init__(self, contrast: Tuple[float] = (0.5, 1.5), saturation: Tuple[float] = (0.5, 1.5), def __init__(
hue: Tuple[float] = (-0.05, 0.05), brightness: Tuple[float] = (0.875, 1.125), p: float = 0.5): self,
contrast: Tuple[float] = (0.5, 1.5),
saturation: Tuple[float] = (0.5, 1.5),
hue: Tuple[float] = (-0.05, 0.05),
brightness: Tuple[float] = (0.875, 1.125),
p: float = 0.5,
):
super().__init__() super().__init__()
self._brightness = T.ColorJitter(brightness=brightness) self._brightness = T.ColorJitter(brightness=brightness)
self._contrast = T.ColorJitter(contrast=contrast) self._contrast = T.ColorJitter(contrast=contrast)
...@@ -216,11 +239,12 @@ class RandomPhotometricDistort(nn.Module): ...@@ -216,11 +239,12 @@ class RandomPhotometricDistort(nn.Module):
self._saturation = T.ColorJitter(saturation=saturation) self._saturation = T.ColorJitter(saturation=saturation)
self.p = p self.p = p
def forward(self, image: Tensor, def forward(
target: Optional[Dict[str, Tensor]] = None) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]: self, image: Tensor, target: Optional[Dict[str, Tensor]] = None
) -> Tuple[Tensor, Optional[Dict[str, Tensor]]]:
if isinstance(image, torch.Tensor): if isinstance(image, torch.Tensor):
if image.ndimension() not in {2, 3}: if image.ndimension() not in {2, 3}:
raise ValueError('image should be 2/3 dimensional. Got {} dimensions.'.format(image.ndimension())) raise ValueError("image should be 2/3 dimensional. Got {} dimensions.".format(image.ndimension()))
elif image.ndimension() == 2: elif image.ndimension() == 2:
image = image.unsqueeze(0) image = image.unsqueeze(0)
......
from collections import defaultdict, deque
import datetime import datetime
import errno import errno
import os import os
import time import time
from collections import defaultdict, deque
import torch import torch
import torch.distributed as dist import torch.distributed as dist
...@@ -32,7 +32,7 @@ class SmoothedValue(object): ...@@ -32,7 +32,7 @@ class SmoothedValue(object):
""" """
if not is_dist_avail_and_initialized(): if not is_dist_avail_and_initialized():
return return
t = torch.tensor([self.count, self.total], dtype=torch.float64, device='cuda') t = torch.tensor([self.count, self.total], dtype=torch.float64, device="cuda")
dist.barrier() dist.barrier()
dist.all_reduce(t) dist.all_reduce(t)
t = t.tolist() t = t.tolist()
...@@ -63,11 +63,8 @@ class SmoothedValue(object): ...@@ -63,11 +63,8 @@ class SmoothedValue(object):
def __str__(self): def __str__(self):
return self.fmt.format( return self.fmt.format(
median=self.median, median=self.median, avg=self.avg, global_avg=self.global_avg, max=self.max, value=self.value
avg=self.avg, )
global_avg=self.global_avg,
max=self.max,
value=self.value)
def all_gather(data): def all_gather(data):
...@@ -130,15 +127,12 @@ class MetricLogger(object): ...@@ -130,15 +127,12 @@ class MetricLogger(object):
return self.meters[attr] return self.meters[attr]
if attr in self.__dict__: if attr in self.__dict__:
return self.__dict__[attr] return self.__dict__[attr]
raise AttributeError("'{}' object has no attribute '{}'".format( raise AttributeError("'{}' object has no attribute '{}'".format(type(self).__name__, attr))
type(self).__name__, attr))
def __str__(self): def __str__(self):
loss_str = [] loss_str = []
for name, meter in self.meters.items(): for name, meter in self.meters.items():
loss_str.append( loss_str.append("{}: {}".format(name, str(meter)))
"{}: {}".format(name, str(meter))
)
return self.delimiter.join(loss_str) return self.delimiter.join(loss_str)
def synchronize_between_processes(self): def synchronize_between_processes(self):
...@@ -151,31 +145,28 @@ class MetricLogger(object): ...@@ -151,31 +145,28 @@ class MetricLogger(object):
def log_every(self, iterable, print_freq, header=None): def log_every(self, iterable, print_freq, header=None):
i = 0 i = 0
if not header: if not header:
header = '' header = ""
start_time = time.time() start_time = time.time()
end = time.time() end = time.time()
iter_time = SmoothedValue(fmt='{avg:.4f}') iter_time = SmoothedValue(fmt="{avg:.4f}")
data_time = SmoothedValue(fmt='{avg:.4f}') data_time = SmoothedValue(fmt="{avg:.4f}")
space_fmt = ':' + str(len(str(len(iterable)))) + 'd' space_fmt = ":" + str(len(str(len(iterable)))) + "d"
if torch.cuda.is_available(): if torch.cuda.is_available():
log_msg = self.delimiter.join([ log_msg = self.delimiter.join(
header, [
'[{0' + space_fmt + '}/{1}]', header,
'eta: {eta}', "[{0" + space_fmt + "}/{1}]",
'{meters}', "eta: {eta}",
'time: {time}', "{meters}",
'data: {data}', "time: {time}",
'max mem: {memory:.0f}' "data: {data}",
]) "max mem: {memory:.0f}",
]
)
else: else:
log_msg = self.delimiter.join([ log_msg = self.delimiter.join(
header, [header, "[{0" + space_fmt + "}/{1}]", "eta: {eta}", "{meters}", "time: {time}", "data: {data}"]
'[{0' + space_fmt + '}/{1}]', )
'eta: {eta}',
'{meters}',
'time: {time}',
'data: {data}'
])
MB = 1024.0 * 1024.0 MB = 1024.0 * 1024.0
for obj in iterable: for obj in iterable:
data_time.update(time.time() - end) data_time.update(time.time() - end)
...@@ -185,22 +176,28 @@ class MetricLogger(object): ...@@ -185,22 +176,28 @@ class MetricLogger(object):
eta_seconds = iter_time.global_avg * (len(iterable) - i) eta_seconds = iter_time.global_avg * (len(iterable) - i)
eta_string = str(datetime.timedelta(seconds=int(eta_seconds))) eta_string = str(datetime.timedelta(seconds=int(eta_seconds)))
if torch.cuda.is_available(): if torch.cuda.is_available():
print(log_msg.format( print(
i, len(iterable), eta=eta_string, log_msg.format(
meters=str(self), i,
time=str(iter_time), data=str(data_time), len(iterable),
memory=torch.cuda.max_memory_allocated() / MB)) eta=eta_string,
meters=str(self),
time=str(iter_time),
data=str(data_time),
memory=torch.cuda.max_memory_allocated() / MB,
)
)
else: else:
print(log_msg.format( print(
i, len(iterable), eta=eta_string, log_msg.format(
meters=str(self), i, len(iterable), eta=eta_string, meters=str(self), time=str(iter_time), data=str(data_time)
time=str(iter_time), data=str(data_time))) )
)
i += 1 i += 1
end = time.time() end = time.time()
total_time = time.time() - start_time total_time = time.time() - start_time
total_time_str = str(datetime.timedelta(seconds=int(total_time))) total_time_str = str(datetime.timedelta(seconds=int(total_time)))
print('{} Total time: {} ({:.4f} s / it)'.format( print("{} Total time: {} ({:.4f} s / it)".format(header, total_time_str, total_time / len(iterable)))
header, total_time_str, total_time / len(iterable)))
def collate_fn(batch): def collate_fn(batch):
...@@ -220,10 +217,11 @@ def setup_for_distributed(is_master): ...@@ -220,10 +217,11 @@ def setup_for_distributed(is_master):
This function disables printing when not in master process This function disables printing when not in master process
""" """
import builtins as __builtin__ import builtins as __builtin__
builtin_print = __builtin__.print builtin_print = __builtin__.print
def print(*args, **kwargs): def print(*args, **kwargs):
force = kwargs.pop('force', False) force = kwargs.pop("force", False)
if is_master or force: if is_master or force:
builtin_print(*args, **kwargs) builtin_print(*args, **kwargs)
...@@ -260,25 +258,25 @@ def save_on_master(*args, **kwargs): ...@@ -260,25 +258,25 @@ def save_on_master(*args, **kwargs):
def init_distributed_mode(args): def init_distributed_mode(args):
if 'RANK' in os.environ and 'WORLD_SIZE' in os.environ: if "RANK" in os.environ and "WORLD_SIZE" in os.environ:
args.rank = int(os.environ["RANK"]) args.rank = int(os.environ["RANK"])
args.world_size = int(os.environ['WORLD_SIZE']) args.world_size = int(os.environ["WORLD_SIZE"])
args.gpu = int(os.environ['LOCAL_RANK']) args.gpu = int(os.environ["LOCAL_RANK"])
elif 'SLURM_PROCID' in os.environ: elif "SLURM_PROCID" in os.environ:
args.rank = int(os.environ['SLURM_PROCID']) args.rank = int(os.environ["SLURM_PROCID"])
args.gpu = args.rank % torch.cuda.device_count() args.gpu = args.rank % torch.cuda.device_count()
else: else:
print('Not using distributed mode') print("Not using distributed mode")
args.distributed = False args.distributed = False
return return
args.distributed = True args.distributed = True
torch.cuda.set_device(args.gpu) torch.cuda.set_device(args.gpu)
args.dist_backend = 'nccl' args.dist_backend = "nccl"
print('| distributed init (rank {}): {}'.format( print("| distributed init (rank {}): {}".format(args.rank, args.dist_url), flush=True)
args.rank, args.dist_url), flush=True) torch.distributed.init_process_group(
torch.distributed.init_process_group(backend=args.dist_backend, init_method=args.dist_url, backend=args.dist_backend, init_method=args.dist_url, world_size=args.world_size, rank=args.rank
world_size=args.world_size, rank=args.rank) )
torch.distributed.barrier() torch.distributed.barrier()
setup_for_distributed(args.rank == 0) setup_for_distributed(args.rank == 0)
import copy import copy
import os
import torch import torch
import torch.utils.data import torch.utils.data
import torchvision import torchvision
from PIL import Image from PIL import Image
import os
from pycocotools import mask as coco_mask from pycocotools import mask as coco_mask
from transforms import Compose from transforms import Compose
...@@ -90,14 +88,9 @@ def get_coco(root, image_set, transforms): ...@@ -90,14 +88,9 @@ def get_coco(root, image_set, transforms):
"val": ("val2017", os.path.join("annotations", "instances_val2017.json")), "val": ("val2017", os.path.join("annotations", "instances_val2017.json")),
# "train": ("val2017", os.path.join("annotations", "instances_val2017.json")) # "train": ("val2017", os.path.join("annotations", "instances_val2017.json"))
} }
CAT_LIST = [0, 5, 2, 16, 9, 44, 6, 3, 17, 62, 21, 67, 18, 19, 4, CAT_LIST = [0, 5, 2, 16, 9, 44, 6, 3, 17, 62, 21, 67, 18, 19, 4, 1, 64, 20, 63, 7, 72]
1, 64, 20, 63, 7, 72]
transforms = Compose([FilterAndRemapCocoCategories(CAT_LIST, remap=True), ConvertCocoPolysToMask(), transforms])
transforms = Compose([
FilterAndRemapCocoCategories(CAT_LIST, remap=True),
ConvertCocoPolysToMask(),
transforms
])
img_folder, ann_file = PATHS[image_set] img_folder, ann_file = PATHS[image_set]
img_folder = os.path.join(root, img_folder) img_folder = os.path.join(root, img_folder)
......
import torch import torch
import transforms as T import transforms as T
...@@ -11,12 +10,14 @@ class SegmentationPresetTrain: ...@@ -11,12 +10,14 @@ class SegmentationPresetTrain:
trans = [T.RandomResize(min_size, max_size)] trans = [T.RandomResize(min_size, max_size)]
if hflip_prob > 0: if hflip_prob > 0:
trans.append(T.RandomHorizontalFlip(hflip_prob)) trans.append(T.RandomHorizontalFlip(hflip_prob))
trans.extend([ trans.extend(
T.RandomCrop(crop_size), [
T.PILToTensor(), T.RandomCrop(crop_size),
T.ConvertImageDtype(torch.float), T.PILToTensor(),
T.Normalize(mean=mean, std=std), T.ConvertImageDtype(torch.float),
]) T.Normalize(mean=mean, std=std),
]
)
self.transforms = T.Compose(trans) self.transforms = T.Compose(trans)
def __call__(self, img, target): def __call__(self, img, target):
...@@ -25,12 +26,14 @@ class SegmentationPresetTrain: ...@@ -25,12 +26,14 @@ class SegmentationPresetTrain:
class SegmentationPresetEval: class SegmentationPresetEval:
def __init__(self, base_size, mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225)): def __init__(self, base_size, mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225)):
self.transforms = T.Compose([ self.transforms = T.Compose(
T.RandomResize(base_size, base_size), [
T.PILToTensor(), T.RandomResize(base_size, base_size),
T.ConvertImageDtype(torch.float), T.PILToTensor(),
T.Normalize(mean=mean, std=std), T.ConvertImageDtype(torch.float),
]) T.Normalize(mean=mean, std=std),
]
)
def __call__(self, img, target): def __call__(self, img, target):
return self.transforms(img, target) return self.transforms(img, target)
...@@ -2,23 +2,23 @@ import datetime ...@@ -2,23 +2,23 @@ import datetime
import os import os
import time import time
import presets
import torch import torch
import torch.utils.data import torch.utils.data
from torch import nn
import torchvision import torchvision
from coco_utils import get_coco
import presets
import utils import utils
from coco_utils import get_coco
from torch import nn
def get_dataset(dir_path, name, image_set, transform): def get_dataset(dir_path, name, image_set, transform):
def sbd(*args, **kwargs): def sbd(*args, **kwargs):
return torchvision.datasets.SBDataset(*args, mode='segmentation', **kwargs) return torchvision.datasets.SBDataset(*args, mode="segmentation", **kwargs)
paths = { paths = {
"voc": (dir_path, torchvision.datasets.VOCSegmentation, 21), "voc": (dir_path, torchvision.datasets.VOCSegmentation, 21),
"voc_aug": (dir_path, sbd, 21), "voc_aug": (dir_path, sbd, 21),
"coco": (dir_path, get_coco, 21) "coco": (dir_path, get_coco, 21),
} }
p, ds_fn, num_classes = paths[name] p, ds_fn, num_classes = paths[name]
...@@ -39,21 +39,21 @@ def criterion(inputs, target): ...@@ -39,21 +39,21 @@ def criterion(inputs, target):
losses[name] = nn.functional.cross_entropy(x, target, ignore_index=255) losses[name] = nn.functional.cross_entropy(x, target, ignore_index=255)
if len(losses) == 1: if len(losses) == 1:
return losses['out'] return losses["out"]
return losses['out'] + 0.5 * losses['aux'] return losses["out"] + 0.5 * losses["aux"]
def evaluate(model, data_loader, device, num_classes): def evaluate(model, data_loader, device, num_classes):
model.eval() model.eval()
confmat = utils.ConfusionMatrix(num_classes) confmat = utils.ConfusionMatrix(num_classes)
metric_logger = utils.MetricLogger(delimiter=" ") metric_logger = utils.MetricLogger(delimiter=" ")
header = 'Test:' header = "Test:"
with torch.no_grad(): with torch.no_grad():
for image, target in metric_logger.log_every(data_loader, 100, header): for image, target in metric_logger.log_every(data_loader, 100, header):
image, target = image.to(device), target.to(device) image, target = image.to(device), target.to(device)
output = model(image) output = model(image)
output = output['out'] output = output["out"]
confmat.update(target.flatten(), output.argmax(1).flatten()) confmat.update(target.flatten(), output.argmax(1).flatten())
...@@ -65,8 +65,8 @@ def evaluate(model, data_loader, device, num_classes): ...@@ -65,8 +65,8 @@ def evaluate(model, data_loader, device, num_classes):
def train_one_epoch(model, criterion, optimizer, data_loader, lr_scheduler, device, epoch, print_freq): def train_one_epoch(model, criterion, optimizer, data_loader, lr_scheduler, device, epoch, print_freq):
model.train() model.train()
metric_logger = utils.MetricLogger(delimiter=" ") metric_logger = utils.MetricLogger(delimiter=" ")
metric_logger.add_meter('lr', utils.SmoothedValue(window_size=1, fmt='{value}')) metric_logger.add_meter("lr", utils.SmoothedValue(window_size=1, fmt="{value}"))
header = 'Epoch: [{}]'.format(epoch) header = "Epoch: [{}]".format(epoch)
for image, target in metric_logger.log_every(data_loader, print_freq, header): for image, target in metric_logger.log_every(data_loader, print_freq, header):
image, target = image.to(device), target.to(device) image, target = image.to(device), target.to(device)
output = model(image) output = model(image)
...@@ -101,18 +101,21 @@ def main(args): ...@@ -101,18 +101,21 @@ def main(args):
test_sampler = torch.utils.data.SequentialSampler(dataset_test) test_sampler = torch.utils.data.SequentialSampler(dataset_test)
data_loader = torch.utils.data.DataLoader( data_loader = torch.utils.data.DataLoader(
dataset, batch_size=args.batch_size, dataset,
sampler=train_sampler, num_workers=args.workers, batch_size=args.batch_size,
collate_fn=utils.collate_fn, drop_last=True) sampler=train_sampler,
num_workers=args.workers,
collate_fn=utils.collate_fn,
drop_last=True,
)
data_loader_test = torch.utils.data.DataLoader( data_loader_test = torch.utils.data.DataLoader(
dataset_test, batch_size=1, dataset_test, batch_size=1, sampler=test_sampler, num_workers=args.workers, collate_fn=utils.collate_fn
sampler=test_sampler, num_workers=args.workers, )
collate_fn=utils.collate_fn)
model = torchvision.models.segmentation.__dict__[args.model](num_classes=num_classes, model = torchvision.models.segmentation.__dict__[args.model](
aux_loss=args.aux_loss, num_classes=num_classes, aux_loss=args.aux_loss, pretrained=args.pretrained
pretrained=args.pretrained) )
model.to(device) model.to(device)
if args.distributed: if args.distributed:
model = torch.nn.SyncBatchNorm.convert_sync_batchnorm(model) model = torch.nn.SyncBatchNorm.convert_sync_batchnorm(model)
...@@ -129,42 +132,42 @@ def main(args): ...@@ -129,42 +132,42 @@ def main(args):
if args.aux_loss: if args.aux_loss:
params = [p for p in model_without_ddp.aux_classifier.parameters() if p.requires_grad] params = [p for p in model_without_ddp.aux_classifier.parameters() if p.requires_grad]
params_to_optimize.append({"params": params, "lr": args.lr * 10}) params_to_optimize.append({"params": params, "lr": args.lr * 10})
optimizer = torch.optim.SGD( optimizer = torch.optim.SGD(params_to_optimize, lr=args.lr, momentum=args.momentum, weight_decay=args.weight_decay)
params_to_optimize,
lr=args.lr, momentum=args.momentum, weight_decay=args.weight_decay)
iters_per_epoch = len(data_loader) iters_per_epoch = len(data_loader)
main_lr_scheduler = torch.optim.lr_scheduler.LambdaLR( main_lr_scheduler = torch.optim.lr_scheduler.LambdaLR(
optimizer, optimizer, lambda x: (1 - x / (iters_per_epoch * (args.epochs - args.lr_warmup_epochs))) ** 0.9
lambda x: (1 - x / (iters_per_epoch * (args.epochs - args.lr_warmup_epochs))) ** 0.9) )
if args.lr_warmup_epochs > 0: if args.lr_warmup_epochs > 0:
warmup_iters = iters_per_epoch * args.lr_warmup_epochs warmup_iters = iters_per_epoch * args.lr_warmup_epochs
args.lr_warmup_method = args.lr_warmup_method.lower() args.lr_warmup_method = args.lr_warmup_method.lower()
if args.lr_warmup_method == 'linear': if args.lr_warmup_method == "linear":
warmup_lr_scheduler = torch.optim.lr_scheduler.LinearLR(optimizer, start_factor=args.lr_warmup_decay, warmup_lr_scheduler = torch.optim.lr_scheduler.LinearLR(
total_iters=warmup_iters) optimizer, start_factor=args.lr_warmup_decay, total_iters=warmup_iters
elif args.lr_warmup_method == 'constant': )
warmup_lr_scheduler = torch.optim.lr_scheduler.ConstantLR(optimizer, factor=args.lr_warmup_decay, elif args.lr_warmup_method == "constant":
total_iters=warmup_iters) warmup_lr_scheduler = torch.optim.lr_scheduler.ConstantLR(
optimizer, factor=args.lr_warmup_decay, total_iters=warmup_iters
)
else: else:
raise RuntimeError("Invalid warmup lr method '{}'. Only linear and constant " raise RuntimeError(
"are supported.".format(args.lr_warmup_method)) "Invalid warmup lr method '{}'. Only linear and constant "
"are supported.".format(args.lr_warmup_method)
)
lr_scheduler = torch.optim.lr_scheduler.SequentialLR( lr_scheduler = torch.optim.lr_scheduler.SequentialLR(
optimizer, optimizer, schedulers=[warmup_lr_scheduler, main_lr_scheduler], milestones=[warmup_iters]
schedulers=[warmup_lr_scheduler, main_lr_scheduler],
milestones=[warmup_iters]
) )
else: else:
lr_scheduler = main_lr_scheduler lr_scheduler = main_lr_scheduler
if args.resume: if args.resume:
checkpoint = torch.load(args.resume, map_location='cpu') checkpoint = torch.load(args.resume, map_location="cpu")
model_without_ddp.load_state_dict(checkpoint['model'], strict=not args.test_only) model_without_ddp.load_state_dict(checkpoint["model"], strict=not args.test_only)
if not args.test_only: if not args.test_only:
optimizer.load_state_dict(checkpoint['optimizer']) optimizer.load_state_dict(checkpoint["optimizer"])
lr_scheduler.load_state_dict(checkpoint['lr_scheduler']) lr_scheduler.load_state_dict(checkpoint["lr_scheduler"])
args.start_epoch = checkpoint['epoch'] + 1 args.start_epoch = checkpoint["epoch"] + 1
if args.test_only: if args.test_only:
confmat = evaluate(model, data_loader_test, device=device, num_classes=num_classes) confmat = evaluate(model, data_loader_test, device=device, num_classes=num_classes)
...@@ -179,53 +182,54 @@ def main(args): ...@@ -179,53 +182,54 @@ def main(args):
confmat = evaluate(model, data_loader_test, device=device, num_classes=num_classes) confmat = evaluate(model, data_loader_test, device=device, num_classes=num_classes)
print(confmat) print(confmat)
checkpoint = { checkpoint = {
'model': model_without_ddp.state_dict(), "model": model_without_ddp.state_dict(),
'optimizer': optimizer.state_dict(), "optimizer": optimizer.state_dict(),
'lr_scheduler': lr_scheduler.state_dict(), "lr_scheduler": lr_scheduler.state_dict(),
'epoch': epoch, "epoch": epoch,
'args': args "args": args,
} }
utils.save_on_master( utils.save_on_master(checkpoint, os.path.join(args.output_dir, "model_{}.pth".format(epoch)))
checkpoint, utils.save_on_master(checkpoint, os.path.join(args.output_dir, "checkpoint.pth"))
os.path.join(args.output_dir, 'model_{}.pth'.format(epoch)))
utils.save_on_master(
checkpoint,
os.path.join(args.output_dir, 'checkpoint.pth'))
total_time = time.time() - start_time total_time = time.time() - start_time
total_time_str = str(datetime.timedelta(seconds=int(total_time))) total_time_str = str(datetime.timedelta(seconds=int(total_time)))
print('Training time {}'.format(total_time_str)) print("Training time {}".format(total_time_str))
def get_args_parser(add_help=True): def get_args_parser(add_help=True):
import argparse import argparse
parser = argparse.ArgumentParser(description='PyTorch Segmentation Training', add_help=add_help)
parser = argparse.ArgumentParser(description="PyTorch Segmentation Training", add_help=add_help)
parser.add_argument('--data-path', default='/datasets01/COCO/022719/', help='dataset path')
parser.add_argument('--dataset', default='coco', help='dataset name') parser.add_argument("--data-path", default="/datasets01/COCO/022719/", help="dataset path")
parser.add_argument('--model', default='fcn_resnet101', help='model') parser.add_argument("--dataset", default="coco", help="dataset name")
parser.add_argument('--aux-loss', action='store_true', help='auxiliar loss') parser.add_argument("--model", default="fcn_resnet101", help="model")
parser.add_argument('--device', default='cuda', help='device') parser.add_argument("--aux-loss", action="store_true", help="auxiliar loss")
parser.add_argument('-b', '--batch-size', default=8, type=int) parser.add_argument("--device", default="cuda", help="device")
parser.add_argument('--epochs', default=30, type=int, metavar='N', parser.add_argument("-b", "--batch-size", default=8, type=int)
help='number of total epochs to run') parser.add_argument("--epochs", default=30, type=int, metavar="N", help="number of total epochs to run")
parser.add_argument('-j', '--workers', default=16, type=int, metavar='N', parser.add_argument(
help='number of data loading workers (default: 16)') "-j", "--workers", default=16, type=int, metavar="N", help="number of data loading workers (default: 16)"
parser.add_argument('--lr', default=0.01, type=float, help='initial learning rate') )
parser.add_argument('--momentum', default=0.9, type=float, metavar='M', parser.add_argument("--lr", default=0.01, type=float, help="initial learning rate")
help='momentum') parser.add_argument("--momentum", default=0.9, type=float, metavar="M", help="momentum")
parser.add_argument('--wd', '--weight-decay', default=1e-4, type=float, parser.add_argument(
metavar='W', help='weight decay (default: 1e-4)', "--wd",
dest='weight_decay') "--weight-decay",
parser.add_argument('--lr-warmup-epochs', default=0, type=int, help='the number of epochs to warmup (default: 0)') default=1e-4,
parser.add_argument('--lr-warmup-method', default="linear", type=str, help='the warmup method (default: linear)') type=float,
parser.add_argument('--lr-warmup-decay', default=0.01, type=float, help='the decay for lr') metavar="W",
parser.add_argument('--print-freq', default=10, type=int, help='print frequency') help="weight decay (default: 1e-4)",
parser.add_argument('--output-dir', default='.', help='path where to save') dest="weight_decay",
parser.add_argument('--resume', default='', help='resume from checkpoint') )
parser.add_argument('--start-epoch', default=0, type=int, metavar='N', parser.add_argument("--lr-warmup-epochs", default=0, type=int, help="the number of epochs to warmup (default: 0)")
help='start epoch') parser.add_argument("--lr-warmup-method", default="linear", type=str, help="the warmup method (default: linear)")
parser.add_argument("--lr-warmup-decay", default=0.01, type=float, help="the decay for lr")
parser.add_argument("--print-freq", default=10, type=int, help="print frequency")
parser.add_argument("--output-dir", default=".", help="path where to save")
parser.add_argument("--resume", default="", help="resume from checkpoint")
parser.add_argument("--start-epoch", default=0, type=int, metavar="N", help="start epoch")
parser.add_argument( parser.add_argument(
"--test-only", "--test-only",
dest="test_only", dest="test_only",
...@@ -239,9 +243,8 @@ def get_args_parser(add_help=True): ...@@ -239,9 +243,8 @@ def get_args_parser(add_help=True):
action="store_true", action="store_true",
) )
# distributed training parameters # distributed training parameters
parser.add_argument('--world-size', default=1, type=int, parser.add_argument("--world-size", default=1, type=int, help="number of distributed processes")
help='number of distributed processes') parser.add_argument("--dist-url", default="env://", help="url used to set up distributed training")
parser.add_argument('--dist-url', default='env://', help='url used to set up distributed training')
return parser return parser
......
from collections import defaultdict, deque
import datetime import datetime
import errno
import os
import time import time
from collections import defaultdict, deque
import torch import torch
import torch.distributed as dist import torch.distributed as dist
import errno
import os
class SmoothedValue(object): class SmoothedValue(object):
"""Track a series of values and provide access to smoothed values over a """Track a series of values and provide access to smoothed values over a
...@@ -32,7 +32,7 @@ class SmoothedValue(object): ...@@ -32,7 +32,7 @@ class SmoothedValue(object):
""" """
if not is_dist_avail_and_initialized(): if not is_dist_avail_and_initialized():
return return
t = torch.tensor([self.count, self.total], dtype=torch.float64, device='cuda') t = torch.tensor([self.count, self.total], dtype=torch.float64, device="cuda")
dist.barrier() dist.barrier()
dist.all_reduce(t) dist.all_reduce(t)
t = t.tolist() t = t.tolist()
...@@ -63,11 +63,8 @@ class SmoothedValue(object): ...@@ -63,11 +63,8 @@ class SmoothedValue(object):
def __str__(self): def __str__(self):
return self.fmt.format( return self.fmt.format(
median=self.median, median=self.median, avg=self.avg, global_avg=self.global_avg, max=self.max, value=self.value
avg=self.avg, )
global_avg=self.global_avg,
max=self.max,
value=self.value)
class ConfusionMatrix(object): class ConfusionMatrix(object):
...@@ -82,7 +79,7 @@ class ConfusionMatrix(object): ...@@ -82,7 +79,7 @@ class ConfusionMatrix(object):
with torch.no_grad(): with torch.no_grad():
k = (a >= 0) & (a < n) k = (a >= 0) & (a < n)
inds = n * a[k].to(torch.int64) + b[k] inds = n * a[k].to(torch.int64) + b[k]
self.mat += torch.bincount(inds, minlength=n**2).reshape(n, n) self.mat += torch.bincount(inds, minlength=n ** 2).reshape(n, n)
def reset(self): def reset(self):
self.mat.zero_() self.mat.zero_()
...@@ -104,15 +101,12 @@ class ConfusionMatrix(object): ...@@ -104,15 +101,12 @@ class ConfusionMatrix(object):
def __str__(self): def __str__(self):
acc_global, acc, iu = self.compute() acc_global, acc, iu = self.compute()
return ( return ("global correct: {:.1f}\n" "average row correct: {}\n" "IoU: {}\n" "mean IoU: {:.1f}").format(
'global correct: {:.1f}\n' acc_global.item() * 100,
'average row correct: {}\n' ["{:.1f}".format(i) for i in (acc * 100).tolist()],
'IoU: {}\n' ["{:.1f}".format(i) for i in (iu * 100).tolist()],
'mean IoU: {:.1f}').format( iu.mean().item() * 100,
acc_global.item() * 100, )
['{:.1f}'.format(i) for i in (acc * 100).tolist()],
['{:.1f}'.format(i) for i in (iu * 100).tolist()],
iu.mean().item() * 100)
class MetricLogger(object): class MetricLogger(object):
...@@ -132,15 +126,12 @@ class MetricLogger(object): ...@@ -132,15 +126,12 @@ class MetricLogger(object):
return self.meters[attr] return self.meters[attr]
if attr in self.__dict__: if attr in self.__dict__:
return self.__dict__[attr] return self.__dict__[attr]
raise AttributeError("'{}' object has no attribute '{}'".format( raise AttributeError("'{}' object has no attribute '{}'".format(type(self).__name__, attr))
type(self).__name__, attr))
def __str__(self): def __str__(self):
loss_str = [] loss_str = []
for name, meter in self.meters.items(): for name, meter in self.meters.items():
loss_str.append( loss_str.append("{}: {}".format(name, str(meter)))
"{}: {}".format(name, str(meter))
)
return self.delimiter.join(loss_str) return self.delimiter.join(loss_str)
def synchronize_between_processes(self): def synchronize_between_processes(self):
...@@ -153,31 +144,28 @@ class MetricLogger(object): ...@@ -153,31 +144,28 @@ class MetricLogger(object):
def log_every(self, iterable, print_freq, header=None): def log_every(self, iterable, print_freq, header=None):
i = 0 i = 0
if not header: if not header:
header = '' header = ""
start_time = time.time() start_time = time.time()
end = time.time() end = time.time()
iter_time = SmoothedValue(fmt='{avg:.4f}') iter_time = SmoothedValue(fmt="{avg:.4f}")
data_time = SmoothedValue(fmt='{avg:.4f}') data_time = SmoothedValue(fmt="{avg:.4f}")
space_fmt = ':' + str(len(str(len(iterable)))) + 'd' space_fmt = ":" + str(len(str(len(iterable)))) + "d"
if torch.cuda.is_available(): if torch.cuda.is_available():
log_msg = self.delimiter.join([ log_msg = self.delimiter.join(
header, [
'[{0' + space_fmt + '}/{1}]', header,
'eta: {eta}', "[{0" + space_fmt + "}/{1}]",
'{meters}', "eta: {eta}",
'time: {time}', "{meters}",
'data: {data}', "time: {time}",
'max mem: {memory:.0f}' "data: {data}",
]) "max mem: {memory:.0f}",
]
)
else: else:
log_msg = self.delimiter.join([ log_msg = self.delimiter.join(
header, [header, "[{0" + space_fmt + "}/{1}]", "eta: {eta}", "{meters}", "time: {time}", "data: {data}"]
'[{0' + space_fmt + '}/{1}]', )
'eta: {eta}',
'{meters}',
'time: {time}',
'data: {data}'
])
MB = 1024.0 * 1024.0 MB = 1024.0 * 1024.0
for obj in iterable: for obj in iterable:
data_time.update(time.time() - end) data_time.update(time.time() - end)
...@@ -187,21 +175,28 @@ class MetricLogger(object): ...@@ -187,21 +175,28 @@ class MetricLogger(object):
eta_seconds = iter_time.global_avg * (len(iterable) - i) eta_seconds = iter_time.global_avg * (len(iterable) - i)
eta_string = str(datetime.timedelta(seconds=int(eta_seconds))) eta_string = str(datetime.timedelta(seconds=int(eta_seconds)))
if torch.cuda.is_available(): if torch.cuda.is_available():
print(log_msg.format( print(
i, len(iterable), eta=eta_string, log_msg.format(
meters=str(self), i,
time=str(iter_time), data=str(data_time), len(iterable),
memory=torch.cuda.max_memory_allocated() / MB)) eta=eta_string,
meters=str(self),
time=str(iter_time),
data=str(data_time),
memory=torch.cuda.max_memory_allocated() / MB,
)
)
else: else:
print(log_msg.format( print(
i, len(iterable), eta=eta_string, log_msg.format(
meters=str(self), i, len(iterable), eta=eta_string, meters=str(self), time=str(iter_time), data=str(data_time)
time=str(iter_time), data=str(data_time))) )
)
i += 1 i += 1
end = time.time() end = time.time()
total_time = time.time() - start_time total_time = time.time() - start_time
total_time_str = str(datetime.timedelta(seconds=int(total_time))) total_time_str = str(datetime.timedelta(seconds=int(total_time)))
print('{} Total time: {}'.format(header, total_time_str)) print("{} Total time: {}".format(header, total_time_str))
def cat_list(images, fill_value=0): def cat_list(images, fill_value=0):
...@@ -209,7 +204,7 @@ def cat_list(images, fill_value=0): ...@@ -209,7 +204,7 @@ def cat_list(images, fill_value=0):
batch_shape = (len(images),) + max_size batch_shape = (len(images),) + max_size
batched_imgs = images[0].new(*batch_shape).fill_(fill_value) batched_imgs = images[0].new(*batch_shape).fill_(fill_value)
for img, pad_img in zip(images, batched_imgs): for img, pad_img in zip(images, batched_imgs):
pad_img[..., :img.shape[-2], :img.shape[-1]].copy_(img) pad_img[..., : img.shape[-2], : img.shape[-1]].copy_(img)
return batched_imgs return batched_imgs
...@@ -233,10 +228,11 @@ def setup_for_distributed(is_master): ...@@ -233,10 +228,11 @@ def setup_for_distributed(is_master):
This function disables printing when not in master process This function disables printing when not in master process
""" """
import builtins as __builtin__ import builtins as __builtin__
builtin_print = __builtin__.print builtin_print = __builtin__.print
def print(*args, **kwargs): def print(*args, **kwargs):
force = kwargs.pop('force', False) force = kwargs.pop("force", False)
if is_master or force: if is_master or force:
builtin_print(*args, **kwargs) builtin_print(*args, **kwargs)
...@@ -273,26 +269,26 @@ def save_on_master(*args, **kwargs): ...@@ -273,26 +269,26 @@ def save_on_master(*args, **kwargs):
def init_distributed_mode(args): def init_distributed_mode(args):
if 'RANK' in os.environ and 'WORLD_SIZE' in os.environ: if "RANK" in os.environ and "WORLD_SIZE" in os.environ:
args.rank = int(os.environ["RANK"]) args.rank = int(os.environ["RANK"])
args.world_size = int(os.environ['WORLD_SIZE']) args.world_size = int(os.environ["WORLD_SIZE"])
args.gpu = int(os.environ['LOCAL_RANK']) args.gpu = int(os.environ["LOCAL_RANK"])
elif 'SLURM_PROCID' in os.environ: elif "SLURM_PROCID" in os.environ:
args.rank = int(os.environ['SLURM_PROCID']) args.rank = int(os.environ["SLURM_PROCID"])
args.gpu = args.rank % torch.cuda.device_count() args.gpu = args.rank % torch.cuda.device_count()
elif hasattr(args, "rank"): elif hasattr(args, "rank"):
pass pass
else: else:
print('Not using distributed mode') print("Not using distributed mode")
args.distributed = False args.distributed = False
return return
args.distributed = True args.distributed = True
torch.cuda.set_device(args.gpu) torch.cuda.set_device(args.gpu)
args.dist_backend = 'nccl' args.dist_backend = "nccl"
print('| distributed init (rank {}): {}'.format( print("| distributed init (rank {}): {}".format(args.rank, args.dist_url), flush=True)
args.rank, args.dist_url), flush=True) torch.distributed.init_process_group(
torch.distributed.init_process_group(backend=args.dist_backend, init_method=args.dist_url, backend=args.dist_backend, init_method=args.dist_url, world_size=args.world_size, rank=args.rank
world_size=args.world_size, rank=args.rank) )
setup_for_distributed(args.rank == 0) setup_for_distributed(args.rank == 0)
''' """
Pytorch adaptation of https://omoindrot.github.io/triplet-loss Pytorch adaptation of https://omoindrot.github.io/triplet-loss
https://github.com/omoindrot/tensorflow-triplet-loss https://github.com/omoindrot/tensorflow-triplet-loss
''' """
import torch import torch
import torch.nn as nn import torch.nn as nn
class TripletMarginLoss(nn.Module): class TripletMarginLoss(nn.Module):
def __init__(self, margin=1.0, p=2., mining='batch_all'): def __init__(self, margin=1.0, p=2.0, mining="batch_all"):
super(TripletMarginLoss, self).__init__() super(TripletMarginLoss, self).__init__()
self.margin = margin self.margin = margin
self.p = p self.p = p
self.mining = mining self.mining = mining
if mining == 'batch_all': if mining == "batch_all":
self.loss_fn = batch_all_triplet_loss self.loss_fn = batch_all_triplet_loss
if mining == 'batch_hard': if mining == "batch_hard":
self.loss_fn = batch_hard_triplet_loss self.loss_fn = batch_hard_triplet_loss
def forward(self, embeddings, labels): def forward(self, embeddings, labels):
......
import random
from collections import defaultdict
import torch import torch
from torch.utils.data.sampler import Sampler from torch.utils.data.sampler import Sampler
from collections import defaultdict
import random
def create_groups(groups, k): def create_groups(groups, k):
......
import unittest import unittest
from collections import defaultdict from collections import defaultdict
from torch.utils.data import DataLoader
from torchvision.datasets import FakeData
import torchvision.transforms as transforms import torchvision.transforms as transforms
from sampler import PKSampler from sampler import PKSampler
from torch.utils.data import DataLoader
from torchvision.datasets import FakeData
class Tester(unittest.TestCase): class Tester(unittest.TestCase):
def test_pksampler(self): def test_pksampler(self):
p, k = 16, 4 p, k = 16, 4
...@@ -19,8 +17,7 @@ class Tester(unittest.TestCase): ...@@ -19,8 +17,7 @@ class Tester(unittest.TestCase):
self.assertRaises(AssertionError, PKSampler, targets, p, k) self.assertRaises(AssertionError, PKSampler, targets, p, k)
# Ensure p, k constraints on batch # Ensure p, k constraints on batch
dataset = FakeData(size=1000, num_classes=100, image_size=(3, 1, 1), dataset = FakeData(size=1000, num_classes=100, image_size=(3, 1, 1), transform=transforms.ToTensor())
transform=transforms.ToTensor())
targets = [target.item() for _, target in dataset] targets = [target.item() for _, target in dataset]
sampler = PKSampler(targets, p, k) sampler = PKSampler(targets, p, k)
loader = DataLoader(dataset, batch_size=p * k, sampler=sampler) loader = DataLoader(dataset, batch_size=p * k, sampler=sampler)
...@@ -38,5 +35,5 @@ class Tester(unittest.TestCase): ...@@ -38,5 +35,5 @@ class Tester(unittest.TestCase):
self.assertEqual(bins[b], k) self.assertEqual(bins[b], k)
if __name__ == '__main__': if __name__ == "__main__":
unittest.main() unittest.main()
import os import os
import torch import torch
from torch.optim import Adam
from torch.utils.data import DataLoader
import torchvision.transforms as transforms import torchvision.transforms as transforms
from torchvision.datasets import FashionMNIST
from loss import TripletMarginLoss from loss import TripletMarginLoss
from sampler import PKSampler
from model import EmbeddingNet from model import EmbeddingNet
from sampler import PKSampler
from torch.optim import Adam
from torch.utils.data import DataLoader
from torchvision.datasets import FashionMNIST
def train_epoch(model, optimizer, criterion, data_loader, device, epoch, print_freq): def train_epoch(model, optimizer, criterion, data_loader, device, epoch, print_freq):
...@@ -33,7 +31,7 @@ def train_epoch(model, optimizer, criterion, data_loader, device, epoch, print_f ...@@ -33,7 +31,7 @@ def train_epoch(model, optimizer, criterion, data_loader, device, epoch, print_f
i += 1 i += 1
avg_loss = running_loss / print_freq avg_loss = running_loss / print_freq
avg_trip = 100.0 * running_frac_pos_triplets / print_freq avg_trip = 100.0 * running_frac_pos_triplets / print_freq
print('[{:d}, {:d}] | loss: {:.4f} | % avg hard triplets: {:.2f}%'.format(epoch, i, avg_loss, avg_trip)) print("[{:d}, {:d}] | loss: {:.4f} | % avg hard triplets: {:.2f}%".format(epoch, i, avg_loss, avg_trip))
running_loss = 0 running_loss = 0
running_frac_pos_triplets = 0 running_frac_pos_triplets = 0
...@@ -79,17 +77,17 @@ def evaluate(model, loader, device): ...@@ -79,17 +77,17 @@ def evaluate(model, loader, device):
threshold, accuracy = find_best_threshold(dists, targets, device) threshold, accuracy = find_best_threshold(dists, targets, device)
print('accuracy: {:.3f}%, threshold: {:.2f}'.format(accuracy, threshold)) print("accuracy: {:.3f}%, threshold: {:.2f}".format(accuracy, threshold))
def save(model, epoch, save_dir, file_name): def save(model, epoch, save_dir, file_name):
file_name = 'epoch_' + str(epoch) + '__' + file_name file_name = "epoch_" + str(epoch) + "__" + file_name
save_path = os.path.join(save_dir, file_name) save_path = os.path.join(save_dir, file_name)
torch.save(model.state_dict(), save_path) torch.save(model.state_dict(), save_path)
def main(args): def main(args):
device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu') device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
p = args.labels_per_batch p = args.labels_per_batch
k = args.samples_per_label k = args.samples_per_label
batch_size = p * k batch_size = p * k
...@@ -103,9 +101,9 @@ def main(args): ...@@ -103,9 +101,9 @@ def main(args):
criterion = TripletMarginLoss(margin=args.margin) criterion = TripletMarginLoss(margin=args.margin)
optimizer = Adam(model.parameters(), lr=args.lr) optimizer = Adam(model.parameters(), lr=args.lr)
transform = transforms.Compose([transforms.Lambda(lambda image: image.convert('RGB')), transform = transforms.Compose(
transforms.Resize((224, 224)), [transforms.Lambda(lambda image: image.convert("RGB")), transforms.Resize((224, 224)), transforms.ToTensor()]
transforms.ToTensor()]) )
# Using FMNIST to demonstrate embedding learning using triplet loss. This dataset can # Using FMNIST to demonstrate embedding learning using triplet loss. This dataset can
# be replaced with any classification dataset. # be replaced with any classification dataset.
...@@ -118,48 +116,44 @@ def main(args): ...@@ -118,48 +116,44 @@ def main(args):
# targets attribute with the same format. # targets attribute with the same format.
targets = train_dataset.targets.tolist() targets = train_dataset.targets.tolist()
train_loader = DataLoader(train_dataset, batch_size=batch_size, train_loader = DataLoader(
sampler=PKSampler(targets, p, k), train_dataset, batch_size=batch_size, sampler=PKSampler(targets, p, k), num_workers=args.workers
num_workers=args.workers) )
test_loader = DataLoader(test_dataset, batch_size=args.eval_batch_size, test_loader = DataLoader(test_dataset, batch_size=args.eval_batch_size, shuffle=False, num_workers=args.workers)
shuffle=False,
num_workers=args.workers)
for epoch in range(1, args.epochs + 1): for epoch in range(1, args.epochs + 1):
print('Training...') print("Training...")
train_epoch(model, optimizer, criterion, train_loader, device, epoch, args.print_freq) train_epoch(model, optimizer, criterion, train_loader, device, epoch, args.print_freq)
print('Evaluating...') print("Evaluating...")
evaluate(model, test_loader, device) evaluate(model, test_loader, device)
print('Saving...') print("Saving...")
save(model, epoch, args.save_dir, 'ckpt.pth') save(model, epoch, args.save_dir, "ckpt.pth")
def parse_args(): def parse_args():
import argparse import argparse
parser = argparse.ArgumentParser(description='PyTorch Embedding Learning')
parser = argparse.ArgumentParser(description="PyTorch Embedding Learning")
parser.add_argument('--dataset-dir', default='/tmp/fmnist/',
help='FashionMNIST dataset directory path') parser.add_argument("--dataset-dir", default="/tmp/fmnist/", help="FashionMNIST dataset directory path")
parser.add_argument('-p', '--labels-per-batch', default=8, type=int, parser.add_argument(
help='Number of unique labels/classes per batch') "-p", "--labels-per-batch", default=8, type=int, help="Number of unique labels/classes per batch"
parser.add_argument('-k', '--samples-per-label', default=8, type=int, )
help='Number of samples per label in a batch') parser.add_argument("-k", "--samples-per-label", default=8, type=int, help="Number of samples per label in a batch")
parser.add_argument('--eval-batch-size', default=512, type=int) parser.add_argument("--eval-batch-size", default=512, type=int)
parser.add_argument('--epochs', default=10, type=int, metavar='N', parser.add_argument("--epochs", default=10, type=int, metavar="N", help="Number of training epochs to run")
help='Number of training epochs to run') parser.add_argument("-j", "--workers", default=4, type=int, metavar="N", help="Number of data loading workers")
parser.add_argument('-j', '--workers', default=4, type=int, metavar='N', parser.add_argument("--lr", default=0.0001, type=float, help="Learning rate")
help='Number of data loading workers') parser.add_argument("--margin", default=0.2, type=float, help="Triplet loss margin")
parser.add_argument('--lr', default=0.0001, type=float, help='Learning rate') parser.add_argument("--print-freq", default=20, type=int, help="Print frequency")
parser.add_argument('--margin', default=0.2, type=float, help='Triplet loss margin') parser.add_argument("--save-dir", default=".", help="Model save directory")
parser.add_argument('--print-freq', default=20, type=int, help='Print frequency') parser.add_argument("--resume", default="", help="Resume from checkpoint")
parser.add_argument('--save-dir', default='.', help='Model save directory')
parser.add_argument('--resume', default='', help='Resume from checkpoint')
return parser.parse_args() return parser.parse_args()
if __name__ == '__main__': if __name__ == "__main__":
args = parse_args() args = parse_args()
main(args) main(args)
import torch import torch
from torchvision.transforms import transforms from torchvision.transforms import transforms
from transforms import ConvertBHWCtoBCHW, ConvertBCHWtoCBHW from transforms import ConvertBHWCtoBCHW, ConvertBCHWtoCBHW
class VideoClassificationPresetTrain: class VideoClassificationPresetTrain:
def __init__(self, resize_size, crop_size, mean=(0.43216, 0.394666, 0.37645), std=(0.22803, 0.22145, 0.216989), def __init__(
hflip_prob=0.5): self,
resize_size,
crop_size,
mean=(0.43216, 0.394666, 0.37645),
std=(0.22803, 0.22145, 0.216989),
hflip_prob=0.5,
):
trans = [ trans = [
ConvertBHWCtoBCHW(), ConvertBHWCtoBCHW(),
transforms.ConvertImageDtype(torch.float32), transforms.ConvertImageDtype(torch.float32),
...@@ -14,11 +19,7 @@ class VideoClassificationPresetTrain: ...@@ -14,11 +19,7 @@ class VideoClassificationPresetTrain:
] ]
if hflip_prob > 0: if hflip_prob > 0:
trans.append(transforms.RandomHorizontalFlip(hflip_prob)) trans.append(transforms.RandomHorizontalFlip(hflip_prob))
trans.extend([ trans.extend([transforms.Normalize(mean=mean, std=std), transforms.RandomCrop(crop_size), ConvertBCHWtoCBHW()])
transforms.Normalize(mean=mean, std=std),
transforms.RandomCrop(crop_size),
ConvertBCHWtoCBHW()
])
self.transforms = transforms.Compose(trans) self.transforms = transforms.Compose(trans)
def __call__(self, x): def __call__(self, x):
...@@ -27,14 +28,16 @@ class VideoClassificationPresetTrain: ...@@ -27,14 +28,16 @@ class VideoClassificationPresetTrain:
class VideoClassificationPresetEval: class VideoClassificationPresetEval:
def __init__(self, resize_size, crop_size, mean=(0.43216, 0.394666, 0.37645), std=(0.22803, 0.22145, 0.216989)): def __init__(self, resize_size, crop_size, mean=(0.43216, 0.394666, 0.37645), std=(0.22803, 0.22145, 0.216989)):
self.transforms = transforms.Compose([ self.transforms = transforms.Compose(
ConvertBHWCtoBCHW(), [
transforms.ConvertImageDtype(torch.float32), ConvertBHWCtoBCHW(),
transforms.Resize(resize_size), transforms.ConvertImageDtype(torch.float32),
transforms.Normalize(mean=mean, std=std), transforms.Resize(resize_size),
transforms.CenterCrop(crop_size), transforms.Normalize(mean=mean, std=std),
ConvertBCHWtoCBHW() transforms.CenterCrop(crop_size),
]) ConvertBCHWtoCBHW(),
]
)
def __call__(self, x): def __call__(self, x):
return self.transforms(x) return self.transforms(x)
import datetime import datetime
import os import os
import time import time
import presets
import torch import torch
import torch.utils.data import torch.utils.data
from torch.utils.data.dataloader import default_collate
from torch import nn
import torchvision import torchvision
import torchvision.datasets.video_utils import torchvision.datasets.video_utils
from torchvision.datasets.samplers import DistributedSampler, UniformClipSampler, RandomClipSampler
import presets
import utils import utils
from torch import nn
from torch.utils.data.dataloader import default_collate
from torchvision.datasets.samplers import DistributedSampler, UniformClipSampler, RandomClipSampler
try: try:
from apex import amp from apex import amp
...@@ -21,10 +21,10 @@ except ImportError: ...@@ -21,10 +21,10 @@ except ImportError:
def train_one_epoch(model, criterion, optimizer, lr_scheduler, data_loader, device, epoch, print_freq, apex=False): def train_one_epoch(model, criterion, optimizer, lr_scheduler, data_loader, device, epoch, print_freq, apex=False):
model.train() model.train()
metric_logger = utils.MetricLogger(delimiter=" ") metric_logger = utils.MetricLogger(delimiter=" ")
metric_logger.add_meter('lr', utils.SmoothedValue(window_size=1, fmt='{value}')) metric_logger.add_meter("lr", utils.SmoothedValue(window_size=1, fmt="{value}"))
metric_logger.add_meter('clips/s', utils.SmoothedValue(window_size=10, fmt='{value:.3f}')) metric_logger.add_meter("clips/s", utils.SmoothedValue(window_size=10, fmt="{value:.3f}"))
header = 'Epoch: [{}]'.format(epoch) header = "Epoch: [{}]".format(epoch)
for video, target in metric_logger.log_every(data_loader, print_freq, header): for video, target in metric_logger.log_every(data_loader, print_freq, header):
start_time = time.time() start_time = time.time()
video, target = video.to(device), target.to(device) video, target = video.to(device), target.to(device)
...@@ -42,16 +42,16 @@ def train_one_epoch(model, criterion, optimizer, lr_scheduler, data_loader, devi ...@@ -42,16 +42,16 @@ def train_one_epoch(model, criterion, optimizer, lr_scheduler, data_loader, devi
acc1, acc5 = utils.accuracy(output, target, topk=(1, 5)) acc1, acc5 = utils.accuracy(output, target, topk=(1, 5))
batch_size = video.shape[0] batch_size = video.shape[0]
metric_logger.update(loss=loss.item(), lr=optimizer.param_groups[0]["lr"]) metric_logger.update(loss=loss.item(), lr=optimizer.param_groups[0]["lr"])
metric_logger.meters['acc1'].update(acc1.item(), n=batch_size) metric_logger.meters["acc1"].update(acc1.item(), n=batch_size)
metric_logger.meters['acc5'].update(acc5.item(), n=batch_size) metric_logger.meters["acc5"].update(acc5.item(), n=batch_size)
metric_logger.meters['clips/s'].update(batch_size / (time.time() - start_time)) metric_logger.meters["clips/s"].update(batch_size / (time.time() - start_time))
lr_scheduler.step() lr_scheduler.step()
def evaluate(model, criterion, data_loader, device): def evaluate(model, criterion, data_loader, device):
model.eval() model.eval()
metric_logger = utils.MetricLogger(delimiter=" ") metric_logger = utils.MetricLogger(delimiter=" ")
header = 'Test:' header = "Test:"
with torch.no_grad(): with torch.no_grad():
for video, target in metric_logger.log_every(data_loader, 100, header): for video, target in metric_logger.log_every(data_loader, 100, header):
video = video.to(device, non_blocking=True) video = video.to(device, non_blocking=True)
...@@ -64,18 +64,22 @@ def evaluate(model, criterion, data_loader, device): ...@@ -64,18 +64,22 @@ def evaluate(model, criterion, data_loader, device):
# could have been padded in distributed setup # could have been padded in distributed setup
batch_size = video.shape[0] batch_size = video.shape[0]
metric_logger.update(loss=loss.item()) metric_logger.update(loss=loss.item())
metric_logger.meters['acc1'].update(acc1.item(), n=batch_size) metric_logger.meters["acc1"].update(acc1.item(), n=batch_size)
metric_logger.meters['acc5'].update(acc5.item(), n=batch_size) metric_logger.meters["acc5"].update(acc5.item(), n=batch_size)
# gather the stats from all processes # gather the stats from all processes
metric_logger.synchronize_between_processes() metric_logger.synchronize_between_processes()
print(' * Clip Acc@1 {top1.global_avg:.3f} Clip Acc@5 {top5.global_avg:.3f}' print(
.format(top1=metric_logger.acc1, top5=metric_logger.acc5)) " * Clip Acc@1 {top1.global_avg:.3f} Clip Acc@5 {top5.global_avg:.3f}".format(
top1=metric_logger.acc1, top5=metric_logger.acc5
)
)
return metric_logger.acc1.global_avg return metric_logger.acc1.global_avg
def _get_cache_path(filepath): def _get_cache_path(filepath):
import hashlib import hashlib
h = hashlib.sha1(filepath.encode()).hexdigest() h = hashlib.sha1(filepath.encode()).hexdigest()
cache_path = os.path.join("~", ".torch", "vision", "datasets", "kinetics", h[:10] + ".pt") cache_path = os.path.join("~", ".torch", "vision", "datasets", "kinetics", h[:10] + ".pt")
cache_path = os.path.expanduser(cache_path) cache_path = os.path.expanduser(cache_path)
...@@ -90,8 +94,10 @@ def collate_fn(batch): ...@@ -90,8 +94,10 @@ def collate_fn(batch):
def main(args): def main(args):
if args.apex and amp is None: if args.apex and amp is None:
raise RuntimeError("Failed to import apex. Please install apex from https://www.github.com/nvidia/apex " raise RuntimeError(
"to enable mixed-precision training.") "Failed to import apex. Please install apex from https://www.github.com/nvidia/apex "
"to enable mixed-precision training."
)
if args.output_dir: if args.output_dir:
utils.mkdir(args.output_dir) utils.mkdir(args.output_dir)
...@@ -121,15 +127,17 @@ def main(args): ...@@ -121,15 +127,17 @@ def main(args):
dataset.transform = transform_train dataset.transform = transform_train
else: else:
if args.distributed: if args.distributed:
print("It is recommended to pre-compute the dataset cache " print("It is recommended to pre-compute the dataset cache " "on a single-gpu first, as it will be faster")
"on a single-gpu first, as it will be faster")
dataset = torchvision.datasets.Kinetics400( dataset = torchvision.datasets.Kinetics400(
traindir, traindir,
frames_per_clip=args.clip_len, frames_per_clip=args.clip_len,
step_between_clips=1, step_between_clips=1,
transform=transform_train, transform=transform_train,
frame_rate=15, frame_rate=15,
extensions=('avi', 'mp4', ) extensions=(
"avi",
"mp4",
),
) )
if args.cache_dataset: if args.cache_dataset:
print("Saving dataset_train to {}".format(cache_path)) print("Saving dataset_train to {}".format(cache_path))
...@@ -149,15 +157,17 @@ def main(args): ...@@ -149,15 +157,17 @@ def main(args):
dataset_test.transform = transform_test dataset_test.transform = transform_test
else: else:
if args.distributed: if args.distributed:
print("It is recommended to pre-compute the dataset cache " print("It is recommended to pre-compute the dataset cache " "on a single-gpu first, as it will be faster")
"on a single-gpu first, as it will be faster")
dataset_test = torchvision.datasets.Kinetics400( dataset_test = torchvision.datasets.Kinetics400(
valdir, valdir,
frames_per_clip=args.clip_len, frames_per_clip=args.clip_len,
step_between_clips=1, step_between_clips=1,
transform=transform_test, transform=transform_test,
frame_rate=15, frame_rate=15,
extensions=('avi', 'mp4',) extensions=(
"avi",
"mp4",
),
) )
if args.cache_dataset: if args.cache_dataset:
print("Saving dataset_test to {}".format(cache_path)) print("Saving dataset_test to {}".format(cache_path))
...@@ -172,14 +182,22 @@ def main(args): ...@@ -172,14 +182,22 @@ def main(args):
test_sampler = DistributedSampler(test_sampler) test_sampler = DistributedSampler(test_sampler)
data_loader = torch.utils.data.DataLoader( data_loader = torch.utils.data.DataLoader(
dataset, batch_size=args.batch_size, dataset,
sampler=train_sampler, num_workers=args.workers, batch_size=args.batch_size,
pin_memory=True, collate_fn=collate_fn) sampler=train_sampler,
num_workers=args.workers,
pin_memory=True,
collate_fn=collate_fn,
)
data_loader_test = torch.utils.data.DataLoader( data_loader_test = torch.utils.data.DataLoader(
dataset_test, batch_size=args.batch_size, dataset_test,
sampler=test_sampler, num_workers=args.workers, batch_size=args.batch_size,
pin_memory=True, collate_fn=collate_fn) sampler=test_sampler,
num_workers=args.workers,
pin_memory=True,
collate_fn=collate_fn,
)
print("Creating model") print("Creating model")
model = torchvision.models.video.__dict__[args.model](pretrained=args.pretrained) model = torchvision.models.video.__dict__[args.model](pretrained=args.pretrained)
...@@ -190,13 +208,10 @@ def main(args): ...@@ -190,13 +208,10 @@ def main(args):
criterion = nn.CrossEntropyLoss() criterion = nn.CrossEntropyLoss()
lr = args.lr * args.world_size lr = args.lr * args.world_size
optimizer = torch.optim.SGD( optimizer = torch.optim.SGD(model.parameters(), lr=lr, momentum=args.momentum, weight_decay=args.weight_decay)
model.parameters(), lr=lr, momentum=args.momentum, weight_decay=args.weight_decay)
if args.apex: if args.apex:
model, optimizer = amp.initialize(model, optimizer, model, optimizer = amp.initialize(model, optimizer, opt_level=args.apex_opt_level)
opt_level=args.apex_opt_level
)
# convert scheduler to be per iteration, not per epoch, for warmup that lasts # convert scheduler to be per iteration, not per epoch, for warmup that lasts
# between different epochs # between different epochs
...@@ -207,20 +222,22 @@ def main(args): ...@@ -207,20 +222,22 @@ def main(args):
if args.lr_warmup_epochs > 0: if args.lr_warmup_epochs > 0:
warmup_iters = iters_per_epoch * args.lr_warmup_epochs warmup_iters = iters_per_epoch * args.lr_warmup_epochs
args.lr_warmup_method = args.lr_warmup_method.lower() args.lr_warmup_method = args.lr_warmup_method.lower()
if args.lr_warmup_method == 'linear': if args.lr_warmup_method == "linear":
warmup_lr_scheduler = torch.optim.lr_scheduler.LinearLR(optimizer, start_factor=args.lr_warmup_decay, warmup_lr_scheduler = torch.optim.lr_scheduler.LinearLR(
total_iters=warmup_iters) optimizer, start_factor=args.lr_warmup_decay, total_iters=warmup_iters
elif args.lr_warmup_method == 'constant': )
warmup_lr_scheduler = torch.optim.lr_scheduler.ConstantLR(optimizer, factor=args.lr_warmup_decay, elif args.lr_warmup_method == "constant":
total_iters=warmup_iters) warmup_lr_scheduler = torch.optim.lr_scheduler.ConstantLR(
optimizer, factor=args.lr_warmup_decay, total_iters=warmup_iters
)
else: else:
raise RuntimeError("Invalid warmup lr method '{}'. Only linear and constant " raise RuntimeError(
"are supported.".format(args.lr_warmup_method)) "Invalid warmup lr method '{}'. Only linear and constant "
"are supported.".format(args.lr_warmup_method)
)
lr_scheduler = torch.optim.lr_scheduler.SequentialLR( lr_scheduler = torch.optim.lr_scheduler.SequentialLR(
optimizer, optimizer, schedulers=[warmup_lr_scheduler, main_lr_scheduler], milestones=[warmup_iters]
schedulers=[warmup_lr_scheduler, main_lr_scheduler],
milestones=[warmup_iters]
) )
else: else:
lr_scheduler = main_lr_scheduler lr_scheduler = main_lr_scheduler
...@@ -231,11 +248,11 @@ def main(args): ...@@ -231,11 +248,11 @@ def main(args):
model_without_ddp = model.module model_without_ddp = model.module
if args.resume: if args.resume:
checkpoint = torch.load(args.resume, map_location='cpu') checkpoint = torch.load(args.resume, map_location="cpu")
model_without_ddp.load_state_dict(checkpoint['model']) model_without_ddp.load_state_dict(checkpoint["model"])
optimizer.load_state_dict(checkpoint['optimizer']) optimizer.load_state_dict(checkpoint["optimizer"])
lr_scheduler.load_state_dict(checkpoint['lr_scheduler']) lr_scheduler.load_state_dict(checkpoint["lr_scheduler"])
args.start_epoch = checkpoint['epoch'] + 1 args.start_epoch = checkpoint["epoch"] + 1
if args.test_only: if args.test_only:
evaluate(model, criterion, data_loader_test, device=device) evaluate(model, criterion, data_loader_test, device=device)
...@@ -246,62 +263,65 @@ def main(args): ...@@ -246,62 +263,65 @@ def main(args):
for epoch in range(args.start_epoch, args.epochs): for epoch in range(args.start_epoch, args.epochs):
if args.distributed: if args.distributed:
train_sampler.set_epoch(epoch) train_sampler.set_epoch(epoch)
train_one_epoch(model, criterion, optimizer, lr_scheduler, data_loader, train_one_epoch(
device, epoch, args.print_freq, args.apex) model, criterion, optimizer, lr_scheduler, data_loader, device, epoch, args.print_freq, args.apex
)
evaluate(model, criterion, data_loader_test, device=device) evaluate(model, criterion, data_loader_test, device=device)
if args.output_dir: if args.output_dir:
checkpoint = { checkpoint = {
'model': model_without_ddp.state_dict(), "model": model_without_ddp.state_dict(),
'optimizer': optimizer.state_dict(), "optimizer": optimizer.state_dict(),
'lr_scheduler': lr_scheduler.state_dict(), "lr_scheduler": lr_scheduler.state_dict(),
'epoch': epoch, "epoch": epoch,
'args': args} "args": args,
utils.save_on_master( }
checkpoint, utils.save_on_master(checkpoint, os.path.join(args.output_dir, "model_{}.pth".format(epoch)))
os.path.join(args.output_dir, 'model_{}.pth'.format(epoch))) utils.save_on_master(checkpoint, os.path.join(args.output_dir, "checkpoint.pth"))
utils.save_on_master(
checkpoint,
os.path.join(args.output_dir, 'checkpoint.pth'))
total_time = time.time() - start_time total_time = time.time() - start_time
total_time_str = str(datetime.timedelta(seconds=int(total_time))) total_time_str = str(datetime.timedelta(seconds=int(total_time)))
print('Training time {}'.format(total_time_str)) print("Training time {}".format(total_time_str))
def parse_args(): def parse_args():
import argparse import argparse
parser = argparse.ArgumentParser(description='PyTorch Video Classification Training')
parser = argparse.ArgumentParser(description="PyTorch Video Classification Training")
parser.add_argument('--data-path', default='/datasets01_101/kinetics/070618/', help='dataset')
parser.add_argument('--train-dir', default='train_avi-480p', help='name of train dir') parser.add_argument("--data-path", default="/datasets01_101/kinetics/070618/", help="dataset")
parser.add_argument('--val-dir', default='val_avi-480p', help='name of val dir') parser.add_argument("--train-dir", default="train_avi-480p", help="name of train dir")
parser.add_argument('--model', default='r2plus1d_18', help='model') parser.add_argument("--val-dir", default="val_avi-480p", help="name of val dir")
parser.add_argument('--device', default='cuda', help='device') parser.add_argument("--model", default="r2plus1d_18", help="model")
parser.add_argument('--clip-len', default=16, type=int, metavar='N', parser.add_argument("--device", default="cuda", help="device")
help='number of frames per clip') parser.add_argument("--clip-len", default=16, type=int, metavar="N", help="number of frames per clip")
parser.add_argument('--clips-per-video', default=5, type=int, metavar='N', parser.add_argument(
help='maximum number of clips per video to consider') "--clips-per-video", default=5, type=int, metavar="N", help="maximum number of clips per video to consider"
parser.add_argument('-b', '--batch-size', default=24, type=int) )
parser.add_argument('--epochs', default=45, type=int, metavar='N', parser.add_argument("-b", "--batch-size", default=24, type=int)
help='number of total epochs to run') parser.add_argument("--epochs", default=45, type=int, metavar="N", help="number of total epochs to run")
parser.add_argument('-j', '--workers', default=10, type=int, metavar='N', parser.add_argument(
help='number of data loading workers (default: 10)') "-j", "--workers", default=10, type=int, metavar="N", help="number of data loading workers (default: 10)"
parser.add_argument('--lr', default=0.01, type=float, help='initial learning rate') )
parser.add_argument('--momentum', default=0.9, type=float, metavar='M', parser.add_argument("--lr", default=0.01, type=float, help="initial learning rate")
help='momentum') parser.add_argument("--momentum", default=0.9, type=float, metavar="M", help="momentum")
parser.add_argument('--wd', '--weight-decay', default=1e-4, type=float, parser.add_argument(
metavar='W', help='weight decay (default: 1e-4)', "--wd",
dest='weight_decay') "--weight-decay",
parser.add_argument('--lr-milestones', nargs='+', default=[20, 30, 40], type=int, help='decrease lr on milestones') default=1e-4,
parser.add_argument('--lr-gamma', default=0.1, type=float, help='decrease lr by a factor of lr-gamma') type=float,
parser.add_argument('--lr-warmup-epochs', default=10, type=int, help='the number of epochs to warmup (default: 10)') metavar="W",
parser.add_argument('--lr-warmup-method', default="linear", type=str, help='the warmup method (default: linear)') help="weight decay (default: 1e-4)",
parser.add_argument('--lr-warmup-decay', default=0.001, type=float, help='the decay for lr') dest="weight_decay",
parser.add_argument('--print-freq', default=10, type=int, help='print frequency') )
parser.add_argument('--output-dir', default='.', help='path where to save') parser.add_argument("--lr-milestones", nargs="+", default=[20, 30, 40], type=int, help="decrease lr on milestones")
parser.add_argument('--resume', default='', help='resume from checkpoint') parser.add_argument("--lr-gamma", default=0.1, type=float, help="decrease lr by a factor of lr-gamma")
parser.add_argument('--start-epoch', default=0, type=int, metavar='N', parser.add_argument("--lr-warmup-epochs", default=10, type=int, help="the number of epochs to warmup (default: 10)")
help='start epoch') parser.add_argument("--lr-warmup-method", default="linear", type=str, help="the warmup method (default: linear)")
parser.add_argument("--lr-warmup-decay", default=0.001, type=float, help="the decay for lr")
parser.add_argument("--print-freq", default=10, type=int, help="print frequency")
parser.add_argument("--output-dir", default=".", help="path where to save")
parser.add_argument("--resume", default="", help="resume from checkpoint")
parser.add_argument("--start-epoch", default=0, type=int, metavar="N", help="start epoch")
parser.add_argument( parser.add_argument(
"--cache-dataset", "--cache-dataset",
dest="cache_dataset", dest="cache_dataset",
...@@ -328,18 +348,19 @@ def parse_args(): ...@@ -328,18 +348,19 @@ def parse_args():
) )
# Mixed precision training parameters # Mixed precision training parameters
parser.add_argument('--apex', action='store_true', parser.add_argument("--apex", action="store_true", help="Use apex for mixed precision training")
help='Use apex for mixed precision training') parser.add_argument(
parser.add_argument('--apex-opt-level', default='O1', type=str, "--apex-opt-level",
help='For apex mixed precision training' default="O1",
'O0 for FP32 training, O1 for mixed precision training.' type=str,
'For further detail, see https://github.com/NVIDIA/apex/tree/master/examples/imagenet' help="For apex mixed precision training"
) "O0 for FP32 training, O1 for mixed precision training."
"For further detail, see https://github.com/NVIDIA/apex/tree/master/examples/imagenet",
)
# distributed training parameters # distributed training parameters
parser.add_argument('--world-size', default=1, type=int, parser.add_argument("--world-size", default=1, type=int, help="number of distributed processes")
help='number of distributed processes') parser.add_argument("--dist-url", default="env://", help="url used to set up distributed training")
parser.add_argument('--dist-url', default='env://', help='url used to set up distributed training')
args = parser.parse_args() args = parser.parse_args()
......
...@@ -3,16 +3,14 @@ import torch.nn as nn ...@@ -3,16 +3,14 @@ import torch.nn as nn
class ConvertBHWCtoBCHW(nn.Module): class ConvertBHWCtoBCHW(nn.Module):
"""Convert tensor from (B, H, W, C) to (B, C, H, W) """Convert tensor from (B, H, W, C) to (B, C, H, W)"""
"""
def forward(self, vid: torch.Tensor) -> torch.Tensor: def forward(self, vid: torch.Tensor) -> torch.Tensor:
return vid.permute(0, 3, 1, 2) return vid.permute(0, 3, 1, 2)
class ConvertBCHWtoCBHW(nn.Module): class ConvertBCHWtoCBHW(nn.Module):
"""Convert tensor from (B, C, H, W) to (C, B, H, W) """Convert tensor from (B, C, H, W) to (C, B, H, W)"""
"""
def forward(self, vid: torch.Tensor) -> torch.Tensor: def forward(self, vid: torch.Tensor) -> torch.Tensor:
return vid.permute(1, 0, 2, 3) return vid.permute(1, 0, 2, 3)
from collections import defaultdict, deque
import datetime import datetime
import errno
import os
import time import time
from collections import defaultdict, deque
import torch import torch
import torch.distributed as dist import torch.distributed as dist
import errno
import os
class SmoothedValue(object): class SmoothedValue(object):
"""Track a series of values and provide access to smoothed values over a """Track a series of values and provide access to smoothed values over a
...@@ -32,7 +32,7 @@ class SmoothedValue(object): ...@@ -32,7 +32,7 @@ class SmoothedValue(object):
""" """
if not is_dist_avail_and_initialized(): if not is_dist_avail_and_initialized():
return return
t = torch.tensor([self.count, self.total], dtype=torch.float64, device='cuda') t = torch.tensor([self.count, self.total], dtype=torch.float64, device="cuda")
dist.barrier() dist.barrier()
dist.all_reduce(t) dist.all_reduce(t)
t = t.tolist() t = t.tolist()
...@@ -63,11 +63,8 @@ class SmoothedValue(object): ...@@ -63,11 +63,8 @@ class SmoothedValue(object):
def __str__(self): def __str__(self):
return self.fmt.format( return self.fmt.format(
median=self.median, median=self.median, avg=self.avg, global_avg=self.global_avg, max=self.max, value=self.value
avg=self.avg, )
global_avg=self.global_avg,
max=self.max,
value=self.value)
class MetricLogger(object): class MetricLogger(object):
...@@ -87,15 +84,12 @@ class MetricLogger(object): ...@@ -87,15 +84,12 @@ class MetricLogger(object):
return self.meters[attr] return self.meters[attr]
if attr in self.__dict__: if attr in self.__dict__:
return self.__dict__[attr] return self.__dict__[attr]
raise AttributeError("'{}' object has no attribute '{}'".format( raise AttributeError("'{}' object has no attribute '{}'".format(type(self).__name__, attr))
type(self).__name__, attr))
def __str__(self): def __str__(self):
loss_str = [] loss_str = []
for name, meter in self.meters.items(): for name, meter in self.meters.items():
loss_str.append( loss_str.append("{}: {}".format(name, str(meter)))
"{}: {}".format(name, str(meter))
)
return self.delimiter.join(loss_str) return self.delimiter.join(loss_str)
def synchronize_between_processes(self): def synchronize_between_processes(self):
...@@ -108,31 +102,28 @@ class MetricLogger(object): ...@@ -108,31 +102,28 @@ class MetricLogger(object):
def log_every(self, iterable, print_freq, header=None): def log_every(self, iterable, print_freq, header=None):
i = 0 i = 0
if not header: if not header:
header = '' header = ""
start_time = time.time() start_time = time.time()
end = time.time() end = time.time()
iter_time = SmoothedValue(fmt='{avg:.4f}') iter_time = SmoothedValue(fmt="{avg:.4f}")
data_time = SmoothedValue(fmt='{avg:.4f}') data_time = SmoothedValue(fmt="{avg:.4f}")
space_fmt = ':' + str(len(str(len(iterable)))) + 'd' space_fmt = ":" + str(len(str(len(iterable)))) + "d"
if torch.cuda.is_available(): if torch.cuda.is_available():
log_msg = self.delimiter.join([ log_msg = self.delimiter.join(
header, [
'[{0' + space_fmt + '}/{1}]', header,
'eta: {eta}', "[{0" + space_fmt + "}/{1}]",
'{meters}', "eta: {eta}",
'time: {time}', "{meters}",
'data: {data}', "time: {time}",
'max mem: {memory:.0f}' "data: {data}",
]) "max mem: {memory:.0f}",
]
)
else: else:
log_msg = self.delimiter.join([ log_msg = self.delimiter.join(
header, [header, "[{0" + space_fmt + "}/{1}]", "eta: {eta}", "{meters}", "time: {time}", "data: {data}"]
'[{0' + space_fmt + '}/{1}]', )
'eta: {eta}',
'{meters}',
'time: {time}',
'data: {data}'
])
MB = 1024.0 * 1024.0 MB = 1024.0 * 1024.0
for obj in iterable: for obj in iterable:
data_time.update(time.time() - end) data_time.update(time.time() - end)
...@@ -142,21 +133,28 @@ class MetricLogger(object): ...@@ -142,21 +133,28 @@ class MetricLogger(object):
eta_seconds = iter_time.global_avg * (len(iterable) - i) eta_seconds = iter_time.global_avg * (len(iterable) - i)
eta_string = str(datetime.timedelta(seconds=int(eta_seconds))) eta_string = str(datetime.timedelta(seconds=int(eta_seconds)))
if torch.cuda.is_available(): if torch.cuda.is_available():
print(log_msg.format( print(
i, len(iterable), eta=eta_string, log_msg.format(
meters=str(self), i,
time=str(iter_time), data=str(data_time), len(iterable),
memory=torch.cuda.max_memory_allocated() / MB)) eta=eta_string,
meters=str(self),
time=str(iter_time),
data=str(data_time),
memory=torch.cuda.max_memory_allocated() / MB,
)
)
else: else:
print(log_msg.format( print(
i, len(iterable), eta=eta_string, log_msg.format(
meters=str(self), i, len(iterable), eta=eta_string, meters=str(self), time=str(iter_time), data=str(data_time)
time=str(iter_time), data=str(data_time))) )
)
i += 1 i += 1
end = time.time() end = time.time()
total_time = time.time() - start_time total_time = time.time() - start_time
total_time_str = str(datetime.timedelta(seconds=int(total_time))) total_time_str = str(datetime.timedelta(seconds=int(total_time)))
print('{} Total time: {}'.format(header, total_time_str)) print("{} Total time: {}".format(header, total_time_str))
def accuracy(output, target, topk=(1,)): def accuracy(output, target, topk=(1,)):
...@@ -189,10 +187,11 @@ def setup_for_distributed(is_master): ...@@ -189,10 +187,11 @@ def setup_for_distributed(is_master):
This function disables printing when not in master process This function disables printing when not in master process
""" """
import builtins as __builtin__ import builtins as __builtin__
builtin_print = __builtin__.print builtin_print = __builtin__.print
def print(*args, **kwargs): def print(*args, **kwargs):
force = kwargs.pop('force', False) force = kwargs.pop("force", False)
if is_master or force: if is_master or force:
builtin_print(*args, **kwargs) builtin_print(*args, **kwargs)
...@@ -229,26 +228,26 @@ def save_on_master(*args, **kwargs): ...@@ -229,26 +228,26 @@ def save_on_master(*args, **kwargs):
def init_distributed_mode(args): def init_distributed_mode(args):
if 'RANK' in os.environ and 'WORLD_SIZE' in os.environ: if "RANK" in os.environ and "WORLD_SIZE" in os.environ:
args.rank = int(os.environ["RANK"]) args.rank = int(os.environ["RANK"])
args.world_size = int(os.environ['WORLD_SIZE']) args.world_size = int(os.environ["WORLD_SIZE"])
args.gpu = int(os.environ['LOCAL_RANK']) args.gpu = int(os.environ["LOCAL_RANK"])
elif 'SLURM_PROCID' in os.environ: elif "SLURM_PROCID" in os.environ:
args.rank = int(os.environ['SLURM_PROCID']) args.rank = int(os.environ["SLURM_PROCID"])
args.gpu = args.rank % torch.cuda.device_count() args.gpu = args.rank % torch.cuda.device_count()
elif hasattr(args, "rank"): elif hasattr(args, "rank"):
pass pass
else: else:
print('Not using distributed mode') print("Not using distributed mode")
args.distributed = False args.distributed = False
return return
args.distributed = True args.distributed = True
torch.cuda.set_device(args.gpu) torch.cuda.set_device(args.gpu)
args.dist_backend = 'nccl' args.dist_backend = "nccl"
print('| distributed init (rank {}): {}'.format( print("| distributed init (rank {}): {}".format(args.rank, args.dist_url), flush=True)
args.rank, args.dist_url), flush=True) torch.distributed.init_process_group(
torch.distributed.init_process_group(backend=args.dist_backend, init_method=args.dist_url, backend=args.dist_backend, init_method=args.dist_url, world_size=args.world_size, rank=args.rank
world_size=args.world_size, rank=args.rank) )
setup_for_distributed(args.rank == 0) setup_for_distributed(args.rank == 0)
...@@ -9,7 +9,13 @@ max-line-length = 120 ...@@ -9,7 +9,13 @@ max-line-length = 120
[flake8] [flake8]
max-line-length = 120 max-line-length = 120
ignore = F401,E402,F403,W503,W504,F821 ignore = E203, E402, W503, W504, F821
per-file-ignores =
__init__.py: F401, F403, F405
./hubconf.py: F401
torchvision/models/mobilenet.py: F401, F403
torchvision/models/quantization/mobilenet.py: F401, F403
test/smoke_test.py: F401
exclude = venv exclude = venv
[pydocstyle] [pydocstyle]
......
import os
import io
import re
import sys
from setuptools import setup, find_packages
from pkg_resources import parse_version, get_distribution, DistributionNotFound
import subprocess
import distutils.command.clean import distutils.command.clean
import distutils.spawn import distutils.spawn
from distutils.version import StrictVersion
import glob import glob
import io
import os
import re
import shutil import shutil
import subprocess
import sys
from distutils.version import StrictVersion
import torch import torch
from pkg_resources import parse_version, get_distribution, DistributionNotFound
from setuptools import setup, find_packages
from torch.utils.cpp_extension import BuildExtension, CppExtension, CUDAExtension, CUDA_HOME from torch.utils.cpp_extension import BuildExtension, CppExtension, CUDAExtension, CUDA_HOME
def read(*names, **kwargs): def read(*names, **kwargs):
with io.open( with io.open(os.path.join(os.path.dirname(__file__), *names), encoding=kwargs.get("encoding", "utf8")) as fp:
os.path.join(os.path.dirname(__file__), *names),
encoding=kwargs.get("encoding", "utf8")
) as fp:
return fp.read() return fp.read()
...@@ -32,26 +29,26 @@ def get_dist(pkgname): ...@@ -32,26 +29,26 @@ def get_dist(pkgname):
cwd = os.path.dirname(os.path.abspath(__file__)) cwd = os.path.dirname(os.path.abspath(__file__))
version_txt = os.path.join(cwd, 'version.txt') version_txt = os.path.join(cwd, "version.txt")
with open(version_txt, 'r') as f: with open(version_txt, "r") as f:
version = f.readline().strip() version = f.readline().strip()
sha = 'Unknown' sha = "Unknown"
package_name = 'torchvision' package_name = "torchvision"
try: try:
sha = subprocess.check_output(['git', 'rev-parse', 'HEAD'], cwd=cwd).decode('ascii').strip() sha = subprocess.check_output(["git", "rev-parse", "HEAD"], cwd=cwd).decode("ascii").strip()
except Exception: except Exception:
pass pass
if os.getenv('BUILD_VERSION'): if os.getenv("BUILD_VERSION"):
version = os.getenv('BUILD_VERSION') version = os.getenv("BUILD_VERSION")
elif sha != 'Unknown': elif sha != "Unknown":
version += '+' + sha[:7] version += "+" + sha[:7]
def write_version_file(): def write_version_file():
version_path = os.path.join(cwd, 'torchvision', 'version.py') version_path = os.path.join(cwd, "torchvision", "version.py")
with open(version_path, 'w') as f: with open(version_path, "w") as f:
f.write("__version__ = '{}'\n".format(version)) f.write("__version__ = '{}'\n".format(version))
f.write("git_version = {}\n".format(repr(sha))) f.write("git_version = {}\n".format(repr(sha)))
f.write("from torchvision.extension import _check_cuda_version\n") f.write("from torchvision.extension import _check_cuda_version\n")
...@@ -59,34 +56,34 @@ def write_version_file(): ...@@ -59,34 +56,34 @@ def write_version_file():
f.write(" cuda = _check_cuda_version()\n") f.write(" cuda = _check_cuda_version()\n")
pytorch_dep = 'torch' pytorch_dep = "torch"
if os.getenv('PYTORCH_VERSION'): if os.getenv("PYTORCH_VERSION"):
pytorch_dep += "==" + os.getenv('PYTORCH_VERSION') pytorch_dep += "==" + os.getenv("PYTORCH_VERSION")
requirements = [ requirements = [
'numpy', "numpy",
pytorch_dep, pytorch_dep,
] ]
# Excluding 8.3.0 because of https://github.com/pytorch/vision/issues/4146 # Excluding 8.3.0 because of https://github.com/pytorch/vision/issues/4146
pillow_ver = ' >= 5.3.0, !=8.3.0' pillow_ver = " >= 5.3.0, !=8.3.0"
pillow_req = 'pillow-simd' if get_dist('pillow-simd') is not None else 'pillow' pillow_req = "pillow-simd" if get_dist("pillow-simd") is not None else "pillow"
requirements.append(pillow_req + pillow_ver) requirements.append(pillow_req + pillow_ver)
def find_library(name, vision_include): def find_library(name, vision_include):
this_dir = os.path.dirname(os.path.abspath(__file__)) this_dir = os.path.dirname(os.path.abspath(__file__))
build_prefix = os.environ.get('BUILD_PREFIX', None) build_prefix = os.environ.get("BUILD_PREFIX", None)
is_conda_build = build_prefix is not None is_conda_build = build_prefix is not None
library_found = False library_found = False
conda_installed = False conda_installed = False
lib_folder = None lib_folder = None
include_folder = None include_folder = None
library_header = '{0}.h'.format(name) library_header = "{0}.h".format(name)
# Lookup in TORCHVISION_INCLUDE or in the package file # Lookup in TORCHVISION_INCLUDE or in the package file
package_path = [os.path.join(this_dir, 'torchvision')] package_path = [os.path.join(this_dir, "torchvision")]
for folder in vision_include + package_path: for folder in vision_include + package_path:
candidate_path = os.path.join(folder, library_header) candidate_path = os.path.join(folder, library_header)
library_found = os.path.exists(candidate_path) library_found = os.path.exists(candidate_path)
...@@ -94,67 +91,66 @@ def find_library(name, vision_include): ...@@ -94,67 +91,66 @@ def find_library(name, vision_include):
break break
if not library_found: if not library_found:
print('Running build on conda-build: {0}'.format(is_conda_build)) print("Running build on conda-build: {0}".format(is_conda_build))
if is_conda_build: if is_conda_build:
# Add conda headers/libraries # Add conda headers/libraries
if os.name == 'nt': if os.name == "nt":
build_prefix = os.path.join(build_prefix, 'Library') build_prefix = os.path.join(build_prefix, "Library")
include_folder = os.path.join(build_prefix, 'include') include_folder = os.path.join(build_prefix, "include")
lib_folder = os.path.join(build_prefix, 'lib') lib_folder = os.path.join(build_prefix, "lib")
library_header_path = os.path.join( library_header_path = os.path.join(include_folder, library_header)
include_folder, library_header)
library_found = os.path.isfile(library_header_path) library_found = os.path.isfile(library_header_path)
conda_installed = library_found conda_installed = library_found
else: else:
# Check if using Anaconda to produce wheels # Check if using Anaconda to produce wheels
conda = distutils.spawn.find_executable('conda') conda = distutils.spawn.find_executable("conda")
is_conda = conda is not None is_conda = conda is not None
print('Running build on conda: {0}'.format(is_conda)) print("Running build on conda: {0}".format(is_conda))
if is_conda: if is_conda:
python_executable = sys.executable python_executable = sys.executable
py_folder = os.path.dirname(python_executable) py_folder = os.path.dirname(python_executable)
if os.name == 'nt': if os.name == "nt":
env_path = os.path.join(py_folder, 'Library') env_path = os.path.join(py_folder, "Library")
else: else:
env_path = os.path.dirname(py_folder) env_path = os.path.dirname(py_folder)
lib_folder = os.path.join(env_path, 'lib') lib_folder = os.path.join(env_path, "lib")
include_folder = os.path.join(env_path, 'include') include_folder = os.path.join(env_path, "include")
library_header_path = os.path.join( library_header_path = os.path.join(include_folder, library_header)
include_folder, library_header)
library_found = os.path.isfile(library_header_path) library_found = os.path.isfile(library_header_path)
conda_installed = library_found conda_installed = library_found
if not library_found: if not library_found:
if sys.platform == 'linux': if sys.platform == "linux":
library_found = os.path.exists('/usr/include/{0}'.format( library_found = os.path.exists("/usr/include/{0}".format(library_header))
library_header)) library_found = library_found or os.path.exists("/usr/local/include/{0}".format(library_header))
library_found = library_found or os.path.exists(
'/usr/local/include/{0}'.format(library_header))
return library_found, conda_installed, include_folder, lib_folder return library_found, conda_installed, include_folder, lib_folder
def get_extensions(): def get_extensions():
this_dir = os.path.dirname(os.path.abspath(__file__)) this_dir = os.path.dirname(os.path.abspath(__file__))
extensions_dir = os.path.join(this_dir, 'torchvision', 'csrc') extensions_dir = os.path.join(this_dir, "torchvision", "csrc")
main_file = glob.glob(os.path.join(extensions_dir, '*.cpp')) + glob.glob(os.path.join(extensions_dir, 'ops', main_file = glob.glob(os.path.join(extensions_dir, "*.cpp")) + glob.glob(
'*.cpp')) os.path.join(extensions_dir, "ops", "*.cpp")
)
source_cpu = ( source_cpu = (
glob.glob(os.path.join(extensions_dir, 'ops', 'autograd', '*.cpp')) + glob.glob(os.path.join(extensions_dir, "ops", "autograd", "*.cpp"))
glob.glob(os.path.join(extensions_dir, 'ops', 'cpu', '*.cpp')) + + glob.glob(os.path.join(extensions_dir, "ops", "cpu", "*.cpp"))
glob.glob(os.path.join(extensions_dir, 'ops', 'quantized', 'cpu', '*.cpp')) + glob.glob(os.path.join(extensions_dir, "ops", "quantized", "cpu", "*.cpp"))
) )
is_rocm_pytorch = False is_rocm_pytorch = False
TORCH_MAJOR = int(torch.__version__.split('.')[0]) TORCH_MAJOR = int(torch.__version__.split(".")[0])
TORCH_MINOR = int(torch.__version__.split('.')[1]) TORCH_MINOR = int(torch.__version__.split(".")[1])
if TORCH_MAJOR > 1 or (TORCH_MAJOR == 1 and TORCH_MINOR >= 5): if TORCH_MAJOR > 1 or (TORCH_MAJOR == 1 and TORCH_MINOR >= 5):
from torch.utils.cpp_extension import ROCM_HOME from torch.utils.cpp_extension import ROCM_HOME
is_rocm_pytorch = True if ((torch.version.hip is not None) and (ROCM_HOME is not None)) else False is_rocm_pytorch = True if ((torch.version.hip is not None) and (ROCM_HOME is not None)) else False
if is_rocm_pytorch: if is_rocm_pytorch:
from torch.utils.hipify import hipify_python from torch.utils.hipify import hipify_python
hipify_python.hipify( hipify_python.hipify(
project_directory=this_dir, project_directory=this_dir,
output_directory=this_dir, output_directory=this_dir,
...@@ -162,25 +158,25 @@ def get_extensions(): ...@@ -162,25 +158,25 @@ def get_extensions():
show_detailed=True, show_detailed=True,
is_pytorch_extension=True, is_pytorch_extension=True,
) )
source_cuda = glob.glob(os.path.join(extensions_dir, 'ops', 'hip', '*.hip')) source_cuda = glob.glob(os.path.join(extensions_dir, "ops", "hip", "*.hip"))
# Copy over additional files # Copy over additional files
for file in glob.glob(r"torchvision/csrc/ops/cuda/*.h"): for file in glob.glob(r"torchvision/csrc/ops/cuda/*.h"):
shutil.copy(file, "torchvision/csrc/ops/hip") shutil.copy(file, "torchvision/csrc/ops/hip")
else: else:
source_cuda = glob.glob(os.path.join(extensions_dir, 'ops', 'cuda', '*.cu')) source_cuda = glob.glob(os.path.join(extensions_dir, "ops", "cuda", "*.cu"))
source_cuda += glob.glob(os.path.join(extensions_dir, 'ops', 'autocast', '*.cpp')) source_cuda += glob.glob(os.path.join(extensions_dir, "ops", "autocast", "*.cpp"))
sources = main_file + source_cpu sources = main_file + source_cpu
extension = CppExtension extension = CppExtension
compile_cpp_tests = os.getenv('WITH_CPP_MODELS_TEST', '0') == '1' compile_cpp_tests = os.getenv("WITH_CPP_MODELS_TEST", "0") == "1"
if compile_cpp_tests: if compile_cpp_tests:
test_dir = os.path.join(this_dir, 'test') test_dir = os.path.join(this_dir, "test")
models_dir = os.path.join(this_dir, 'torchvision', 'csrc', 'models') models_dir = os.path.join(this_dir, "torchvision", "csrc", "models")
test_file = glob.glob(os.path.join(test_dir, '*.cpp')) test_file = glob.glob(os.path.join(test_dir, "*.cpp"))
source_models = glob.glob(os.path.join(models_dir, '*.cpp')) source_models = glob.glob(os.path.join(models_dir, "*.cpp"))
test_file = [os.path.join(test_dir, s) for s in test_file] test_file = [os.path.join(test_dir, s) for s in test_file]
source_models = [os.path.join(models_dir, s) for s in source_models] source_models = [os.path.join(models_dir, s) for s in source_models]
...@@ -189,39 +185,38 @@ def get_extensions(): ...@@ -189,39 +185,38 @@ def get_extensions():
define_macros = [] define_macros = []
extra_compile_args = {'cxx': []} extra_compile_args = {"cxx": []}
if (torch.cuda.is_available() and ((CUDA_HOME is not None) or is_rocm_pytorch)) \ if (torch.cuda.is_available() and ((CUDA_HOME is not None) or is_rocm_pytorch)) or os.getenv(
or os.getenv('FORCE_CUDA', '0') == '1': "FORCE_CUDA", "0"
) == "1":
extension = CUDAExtension extension = CUDAExtension
sources += source_cuda sources += source_cuda
if not is_rocm_pytorch: if not is_rocm_pytorch:
define_macros += [('WITH_CUDA', None)] define_macros += [("WITH_CUDA", None)]
nvcc_flags = os.getenv('NVCC_FLAGS', '') nvcc_flags = os.getenv("NVCC_FLAGS", "")
if nvcc_flags == '': if nvcc_flags == "":
nvcc_flags = [] nvcc_flags = []
else: else:
nvcc_flags = nvcc_flags.split(' ') nvcc_flags = nvcc_flags.split(" ")
else: else:
define_macros += [('WITH_HIP', None)] define_macros += [("WITH_HIP", None)]
nvcc_flags = [] nvcc_flags = []
extra_compile_args["nvcc"] = nvcc_flags extra_compile_args["nvcc"] = nvcc_flags
if sys.platform == 'win32': if sys.platform == "win32":
define_macros += [('torchvision_EXPORTS', None)] define_macros += [("torchvision_EXPORTS", None)]
extra_compile_args['cxx'].append('/MP') extra_compile_args["cxx"].append("/MP")
debug_mode = os.getenv('DEBUG', '0') == '1' debug_mode = os.getenv("DEBUG", "0") == "1"
if debug_mode: if debug_mode:
print("Compile in debug mode") print("Compile in debug mode")
extra_compile_args['cxx'].append("-g") extra_compile_args["cxx"].append("-g")
extra_compile_args['cxx'].append("-O0") extra_compile_args["cxx"].append("-O0")
if "nvcc" in extra_compile_args: if "nvcc" in extra_compile_args:
# we have to remove "-OX" and "-g" flag if exists and append # we have to remove "-OX" and "-g" flag if exists and append
nvcc_flags = extra_compile_args["nvcc"] nvcc_flags = extra_compile_args["nvcc"]
extra_compile_args["nvcc"] = [ extra_compile_args["nvcc"] = [f for f in nvcc_flags if not ("-O" in f or "-g" in f)]
f for f in nvcc_flags if not ("-O" in f or "-g" in f)
]
extra_compile_args["nvcc"].append("-O0") extra_compile_args["nvcc"].append("-O0")
extra_compile_args["nvcc"].append("-g") extra_compile_args["nvcc"].append("-g")
...@@ -231,7 +226,7 @@ def get_extensions(): ...@@ -231,7 +226,7 @@ def get_extensions():
ext_modules = [ ext_modules = [
extension( extension(
'torchvision._C', "torchvision._C",
sorted(sources), sorted(sources),
include_dirs=include_dirs, include_dirs=include_dirs,
define_macros=define_macros, define_macros=define_macros,
...@@ -241,7 +236,7 @@ def get_extensions(): ...@@ -241,7 +236,7 @@ def get_extensions():
if compile_cpp_tests: if compile_cpp_tests:
ext_modules.append( ext_modules.append(
extension( extension(
'torchvision._C_tests', "torchvision._C_tests",
tests, tests,
include_dirs=tests_include_dirs, include_dirs=tests_include_dirs,
define_macros=define_macros, define_macros=define_macros,
...@@ -250,12 +245,10 @@ def get_extensions(): ...@@ -250,12 +245,10 @@ def get_extensions():
) )
# ------------------- Torchvision extra extensions ------------------------ # ------------------- Torchvision extra extensions ------------------------
vision_include = os.environ.get('TORCHVISION_INCLUDE', None) vision_include = os.environ.get("TORCHVISION_INCLUDE", None)
vision_library = os.environ.get('TORCHVISION_LIBRARY', None) vision_library = os.environ.get("TORCHVISION_LIBRARY", None)
vision_include = (vision_include.split(os.pathsep) vision_include = vision_include.split(os.pathsep) if vision_include is not None else []
if vision_include is not None else []) vision_library = vision_library.split(os.pathsep) if vision_library is not None else []
vision_library = (vision_library.split(os.pathsep)
if vision_library is not None else [])
include_dirs += vision_include include_dirs += vision_include
library_dirs = vision_library library_dirs = vision_library
...@@ -266,56 +259,49 @@ def get_extensions(): ...@@ -266,56 +259,49 @@ def get_extensions():
image_link_flags = [] image_link_flags = []
# Locating libPNG # Locating libPNG
libpng = distutils.spawn.find_executable('libpng-config') libpng = distutils.spawn.find_executable("libpng-config")
pngfix = distutils.spawn.find_executable('pngfix') pngfix = distutils.spawn.find_executable("pngfix")
png_found = libpng is not None or pngfix is not None png_found = libpng is not None or pngfix is not None
print('PNG found: {0}'.format(png_found)) print("PNG found: {0}".format(png_found))
if png_found: if png_found:
if libpng is not None: if libpng is not None:
# Linux / Mac # Linux / Mac
png_version = subprocess.run([libpng, '--version'], png_version = subprocess.run([libpng, "--version"], stdout=subprocess.PIPE)
stdout=subprocess.PIPE) png_version = png_version.stdout.strip().decode("utf-8")
png_version = png_version.stdout.strip().decode('utf-8') print("libpng version: {0}".format(png_version))
print('libpng version: {0}'.format(png_version))
png_version = parse_version(png_version) png_version = parse_version(png_version)
if png_version >= parse_version("1.6.0"): if png_version >= parse_version("1.6.0"):
print('Building torchvision with PNG image support') print("Building torchvision with PNG image support")
png_lib = subprocess.run([libpng, '--libdir'], png_lib = subprocess.run([libpng, "--libdir"], stdout=subprocess.PIPE)
stdout=subprocess.PIPE) png_lib = png_lib.stdout.strip().decode("utf-8")
png_lib = png_lib.stdout.strip().decode('utf-8') if "disabled" not in png_lib:
if 'disabled' not in png_lib:
image_library += [png_lib] image_library += [png_lib]
png_include = subprocess.run([libpng, '--I_opts'], png_include = subprocess.run([libpng, "--I_opts"], stdout=subprocess.PIPE)
stdout=subprocess.PIPE) png_include = png_include.stdout.strip().decode("utf-8")
png_include = png_include.stdout.strip().decode('utf-8') _, png_include = png_include.split("-I")
_, png_include = png_include.split('-I') print("libpng include path: {0}".format(png_include))
print('libpng include path: {0}'.format(png_include))
image_include += [png_include] image_include += [png_include]
image_link_flags.append('png') image_link_flags.append("png")
else: else:
print('libpng installed version is less than 1.6.0, ' print("libpng installed version is less than 1.6.0, " "disabling PNG support")
'disabling PNG support')
png_found = False png_found = False
else: else:
# Windows # Windows
png_lib = os.path.join( png_lib = os.path.join(os.path.dirname(os.path.dirname(pngfix)), "lib")
os.path.dirname(os.path.dirname(pngfix)), 'lib') png_include = os.path.join(os.path.dirname(os.path.dirname(pngfix)), "include", "libpng16")
png_include = os.path.join(os.path.dirname(
os.path.dirname(pngfix)), 'include', 'libpng16')
image_library += [png_lib] image_library += [png_lib]
image_include += [png_include] image_include += [png_include]
image_link_flags.append('libpng') image_link_flags.append("libpng")
# Locating libjpeg # Locating libjpeg
(jpeg_found, jpeg_conda, (jpeg_found, jpeg_conda, jpeg_include, jpeg_lib) = find_library("jpeglib", vision_include)
jpeg_include, jpeg_lib) = find_library('jpeglib', vision_include)
print('JPEG found: {0}'.format(jpeg_found)) print("JPEG found: {0}".format(jpeg_found))
image_macros += [('PNG_FOUND', str(int(png_found)))] image_macros += [("PNG_FOUND", str(int(png_found)))]
image_macros += [('JPEG_FOUND', str(int(jpeg_found)))] image_macros += [("JPEG_FOUND", str(int(jpeg_found)))]
if jpeg_found: if jpeg_found:
print('Building torchvision with JPEG image support') print("Building torchvision with JPEG image support")
image_link_flags.append('jpeg') image_link_flags.append("jpeg")
if jpeg_conda: if jpeg_conda:
image_library += [jpeg_lib] image_library += [jpeg_lib]
image_include += [jpeg_include] image_include += [jpeg_include]
...@@ -323,80 +309,71 @@ def get_extensions(): ...@@ -323,80 +309,71 @@ def get_extensions():
# Locating nvjpeg # Locating nvjpeg
# Should be included in CUDA_HOME for CUDA >= 10.1, which is the minimum version we have in the CI # Should be included in CUDA_HOME for CUDA >= 10.1, which is the minimum version we have in the CI
nvjpeg_found = ( nvjpeg_found = (
extension is CUDAExtension and extension is CUDAExtension
CUDA_HOME is not None and and CUDA_HOME is not None
os.path.exists(os.path.join(CUDA_HOME, 'include', 'nvjpeg.h')) and os.path.exists(os.path.join(CUDA_HOME, "include", "nvjpeg.h"))
) )
print('NVJPEG found: {0}'.format(nvjpeg_found)) print("NVJPEG found: {0}".format(nvjpeg_found))
image_macros += [('NVJPEG_FOUND', str(int(nvjpeg_found)))] image_macros += [("NVJPEG_FOUND", str(int(nvjpeg_found)))]
if nvjpeg_found: if nvjpeg_found:
print('Building torchvision with NVJPEG image support') print("Building torchvision with NVJPEG image support")
image_link_flags.append('nvjpeg') image_link_flags.append("nvjpeg")
image_path = os.path.join(extensions_dir, 'io', 'image') image_path = os.path.join(extensions_dir, "io", "image")
image_src = (glob.glob(os.path.join(image_path, '*.cpp')) + glob.glob(os.path.join(image_path, 'cpu', '*.cpp')) image_src = (
+ glob.glob(os.path.join(image_path, 'cuda', '*.cpp'))) glob.glob(os.path.join(image_path, "*.cpp"))
+ glob.glob(os.path.join(image_path, "cpu", "*.cpp"))
+ glob.glob(os.path.join(image_path, "cuda", "*.cpp"))
)
if png_found or jpeg_found: if png_found or jpeg_found:
ext_modules.append(extension( ext_modules.append(
'torchvision.image', extension(
image_src, "torchvision.image",
include_dirs=image_include + include_dirs + [image_path], image_src,
library_dirs=image_library + library_dirs, include_dirs=image_include + include_dirs + [image_path],
define_macros=image_macros, library_dirs=image_library + library_dirs,
libraries=image_link_flags, define_macros=image_macros,
extra_compile_args=extra_compile_args libraries=image_link_flags,
)) extra_compile_args=extra_compile_args,
)
ffmpeg_exe = distutils.spawn.find_executable('ffmpeg') )
ffmpeg_exe = distutils.spawn.find_executable("ffmpeg")
has_ffmpeg = ffmpeg_exe is not None has_ffmpeg = ffmpeg_exe is not None
# FIXME: Building torchvision with ffmpeg on MacOS or with Python 3.9 # FIXME: Building torchvision with ffmpeg on MacOS or with Python 3.9
# FIXME: causes crash. See the following GitHub issues for more details. # FIXME: causes crash. See the following GitHub issues for more details.
# FIXME: https://github.com/pytorch/pytorch/issues/65000 # FIXME: https://github.com/pytorch/pytorch/issues/65000
# FIXME: https://github.com/pytorch/vision/issues/3367 # FIXME: https://github.com/pytorch/vision/issues/3367
if sys.platform != 'linux' or ( if sys.platform != "linux" or (sys.version_info.major == 3 and sys.version_info.minor == 9):
sys.version_info.major == 3 and sys.version_info.minor == 9):
has_ffmpeg = False has_ffmpeg = False
if has_ffmpeg: if has_ffmpeg:
try: try:
# This is to check if ffmpeg is installed properly. # This is to check if ffmpeg is installed properly.
subprocess.check_output(["ffmpeg", "-version"]) subprocess.check_output(["ffmpeg", "-version"])
except subprocess.CalledProcessError: except subprocess.CalledProcessError:
print('Error fetching ffmpeg version, ignoring ffmpeg.') print("Error fetching ffmpeg version, ignoring ffmpeg.")
has_ffmpeg = False has_ffmpeg = False
print("FFmpeg found: {}".format(has_ffmpeg)) print("FFmpeg found: {}".format(has_ffmpeg))
if has_ffmpeg: if has_ffmpeg:
ffmpeg_libraries = { ffmpeg_libraries = {"libavcodec", "libavformat", "libavutil", "libswresample", "libswscale"}
'libavcodec',
'libavformat',
'libavutil',
'libswresample',
'libswscale'
}
ffmpeg_bin = os.path.dirname(ffmpeg_exe) ffmpeg_bin = os.path.dirname(ffmpeg_exe)
ffmpeg_root = os.path.dirname(ffmpeg_bin) ffmpeg_root = os.path.dirname(ffmpeg_bin)
ffmpeg_include_dir = os.path.join(ffmpeg_root, 'include') ffmpeg_include_dir = os.path.join(ffmpeg_root, "include")
ffmpeg_library_dir = os.path.join(ffmpeg_root, 'lib') ffmpeg_library_dir = os.path.join(ffmpeg_root, "lib")
gcc = distutils.spawn.find_executable('gcc') gcc = distutils.spawn.find_executable("gcc")
platform_tag = subprocess.run( platform_tag = subprocess.run([gcc, "-print-multiarch"], stdout=subprocess.PIPE)
[gcc, '-print-multiarch'], stdout=subprocess.PIPE) platform_tag = platform_tag.stdout.strip().decode("utf-8")
platform_tag = platform_tag.stdout.strip().decode('utf-8')
if platform_tag: if platform_tag:
# Most probably a Debian-based distribution # Most probably a Debian-based distribution
ffmpeg_include_dir = [ ffmpeg_include_dir = [ffmpeg_include_dir, os.path.join(ffmpeg_include_dir, platform_tag)]
ffmpeg_include_dir, ffmpeg_library_dir = [ffmpeg_library_dir, os.path.join(ffmpeg_library_dir, platform_tag)]
os.path.join(ffmpeg_include_dir, platform_tag)
]
ffmpeg_library_dir = [
ffmpeg_library_dir,
os.path.join(ffmpeg_library_dir, platform_tag)
]
else: else:
ffmpeg_include_dir = [ffmpeg_include_dir] ffmpeg_include_dir = [ffmpeg_include_dir]
ffmpeg_library_dir = [ffmpeg_library_dir] ffmpeg_library_dir = [ffmpeg_library_dir]
...@@ -405,11 +382,11 @@ def get_extensions(): ...@@ -405,11 +382,11 @@ def get_extensions():
for library in ffmpeg_libraries: for library in ffmpeg_libraries:
library_found = False library_found = False
for search_path in ffmpeg_include_dir + include_dirs: for search_path in ffmpeg_include_dir + include_dirs:
full_path = os.path.join(search_path, library, '*.h') full_path = os.path.join(search_path, library, "*.h")
library_found |= len(glob.glob(full_path)) > 0 library_found |= len(glob.glob(full_path)) > 0
if not library_found: if not library_found:
print(f'{library} header files were not found, disabling ffmpeg support') print(f"{library} header files were not found, disabling ffmpeg support")
has_ffmpeg = False has_ffmpeg = False
if has_ffmpeg: if has_ffmpeg:
...@@ -417,22 +394,21 @@ def get_extensions(): ...@@ -417,22 +394,21 @@ def get_extensions():
print("ffmpeg library_dir: {}".format(ffmpeg_library_dir)) print("ffmpeg library_dir: {}".format(ffmpeg_library_dir))
# TorchVision base decoder + video reader # TorchVision base decoder + video reader
video_reader_src_dir = os.path.join(this_dir, 'torchvision', 'csrc', 'io', 'video_reader') video_reader_src_dir = os.path.join(this_dir, "torchvision", "csrc", "io", "video_reader")
video_reader_src = glob.glob(os.path.join(video_reader_src_dir, "*.cpp")) video_reader_src = glob.glob(os.path.join(video_reader_src_dir, "*.cpp"))
base_decoder_src_dir = os.path.join(this_dir, 'torchvision', 'csrc', 'io', 'decoder') base_decoder_src_dir = os.path.join(this_dir, "torchvision", "csrc", "io", "decoder")
base_decoder_src = glob.glob( base_decoder_src = glob.glob(os.path.join(base_decoder_src_dir, "*.cpp"))
os.path.join(base_decoder_src_dir, "*.cpp"))
# Torchvision video API # Torchvision video API
videoapi_src_dir = os.path.join(this_dir, 'torchvision', 'csrc', 'io', 'video') videoapi_src_dir = os.path.join(this_dir, "torchvision", "csrc", "io", "video")
videoapi_src = glob.glob(os.path.join(videoapi_src_dir, "*.cpp")) videoapi_src = glob.glob(os.path.join(videoapi_src_dir, "*.cpp"))
# exclude tests # exclude tests
base_decoder_src = [x for x in base_decoder_src if '_test.cpp' not in x] base_decoder_src = [x for x in base_decoder_src if "_test.cpp" not in x]
combined_src = video_reader_src + base_decoder_src + videoapi_src combined_src = video_reader_src + base_decoder_src + videoapi_src
ext_modules.append( ext_modules.append(
CppExtension( CppExtension(
'torchvision.video_reader', "torchvision.video_reader",
combined_src, combined_src,
include_dirs=[ include_dirs=[
base_decoder_src_dir, base_decoder_src_dir,
...@@ -440,18 +416,18 @@ def get_extensions(): ...@@ -440,18 +416,18 @@ def get_extensions():
videoapi_src_dir, videoapi_src_dir,
extensions_dir, extensions_dir,
*ffmpeg_include_dir, *ffmpeg_include_dir,
*include_dirs *include_dirs,
], ],
library_dirs=ffmpeg_library_dir + library_dirs, library_dirs=ffmpeg_library_dir + library_dirs,
libraries=[ libraries=[
'avcodec', "avcodec",
'avformat', "avformat",
'avutil', "avutil",
'swresample', "swresample",
'swscale', "swscale",
], ],
extra_compile_args=["-std=c++14"] if os.name != 'nt' else ['/std:c++14', '/MP'], extra_compile_args=["-std=c++14"] if os.name != "nt" else ["/std:c++14", "/MP"],
extra_link_args=["-std=c++14" if os.name != 'nt' else '/std:c++14'], extra_link_args=["-std=c++14" if os.name != "nt" else "/std:c++14"],
) )
) )
...@@ -460,9 +436,9 @@ def get_extensions(): ...@@ -460,9 +436,9 @@ def get_extensions():
class clean(distutils.command.clean.clean): class clean(distutils.command.clean.clean):
def run(self): def run(self):
with open('.gitignore', 'r') as f: with open(".gitignore", "r") as f:
ignores = f.read() ignores = f.read()
for wildcard in filter(None, ignores.split('\n')): for wildcard in filter(None, ignores.split("\n")):
for filename in glob.glob(wildcard): for filename in glob.glob(wildcard):
try: try:
os.remove(filename) os.remove(filename)
...@@ -478,25 +454,22 @@ if __name__ == "__main__": ...@@ -478,25 +454,22 @@ if __name__ == "__main__":
write_version_file() write_version_file()
with open('README.rst') as f: with open("README.rst") as f:
readme = f.read() readme = f.read()
setup( setup(
# Metadata # Metadata
name=package_name, name=package_name,
version=version, version=version,
author='PyTorch Core Team', author="PyTorch Core Team",
author_email='soumith@pytorch.org', author_email="soumith@pytorch.org",
url='https://github.com/pytorch/vision', url="https://github.com/pytorch/vision",
description='image and video datasets and models for torch deep learning', description="image and video datasets and models for torch deep learning",
long_description=readme, long_description=readme,
license='BSD', license="BSD",
# Package info # Package info
packages=find_packages(exclude=('test',)), packages=find_packages(exclude=("test",)),
package_data={ package_data={package_name: ["*.dll", "*.dylib", "*.so", "*.categories"]},
package_name: ['*.dll', '*.dylib', '*.so', '*.categories']
},
zip_safe=False, zip_safe=False,
install_requires=requirements, install_requires=requirements,
extras_require={ extras_require={
...@@ -504,7 +477,7 @@ if __name__ == "__main__": ...@@ -504,7 +477,7 @@ if __name__ == "__main__":
}, },
ext_modules=get_extensions(), ext_modules=get_extensions(),
cmdclass={ cmdclass={
'build_ext': BuildExtension.with_options(no_python_abi_suffix=True), "build_ext": BuildExtension.with_options(no_python_abi_suffix=True),
'clean': clean, "clean": clean,
} },
) )
import argparse
import contextlib
import functools
import inspect
import os import os
import random
import shutil import shutil
import sys
import tempfile import tempfile
import contextlib
import unittest import unittest
from collections import OrderedDict
from numbers import Number
import numpy as np
import pytest import pytest
import argparse
import sys
import torch import torch
import __main__ from PIL import Image
import random
import inspect
import functools
from numbers import Number
from torch._six import string_classes from torch._six import string_classes
from collections import OrderedDict
from torchvision import io from torchvision import io
import numpy as np import __main__
from PIL import Image
IN_CIRCLE_CI = os.getenv("CIRCLECI", False) == 'true' IN_CIRCLE_CI = os.getenv("CIRCLECI", False) == "true"
IN_RE_WORKER = os.environ.get("INSIDE_RE_WORKER") is not None IN_RE_WORKER = os.environ.get("INSIDE_RE_WORKER") is not None
IN_FBCODE = os.environ.get("IN_FBCODE_TORCHVISION") == "1" IN_FBCODE = os.environ.get("IN_FBCODE_TORCHVISION") == "1"
CUDA_NOT_AVAILABLE_MSG = 'CUDA device not available' CUDA_NOT_AVAILABLE_MSG = "CUDA device not available"
CIRCLECI_GPU_NO_CUDA_MSG = "We're in a CircleCI GPU machine, and this test doesn't need cuda." CIRCLECI_GPU_NO_CUDA_MSG = "We're in a CircleCI GPU machine, and this test doesn't need cuda."
...@@ -95,7 +95,7 @@ def freeze_rng_state(): ...@@ -95,7 +95,7 @@ def freeze_rng_state():
def cycle_over(objs): def cycle_over(objs):
for idx, obj1 in enumerate(objs): for idx, obj1 in enumerate(objs):
for obj2 in objs[:idx] + objs[idx + 1:]: for obj2 in objs[:idx] + objs[idx + 1 :]:
yield obj1, obj2 yield obj1, obj2
...@@ -117,11 +117,13 @@ def disable_console_output(): ...@@ -117,11 +117,13 @@ def disable_console_output():
def cpu_and_gpu(): def cpu_and_gpu():
import pytest # noqa import pytest # noqa
return ('cpu', pytest.param('cuda', marks=pytest.mark.needs_cuda))
return ("cpu", pytest.param("cuda", marks=pytest.mark.needs_cuda))
def needs_cuda(test_func): def needs_cuda(test_func):
import pytest # noqa import pytest # noqa
return pytest.mark.needs_cuda(test_func) return pytest.mark.needs_cuda(test_func)
...@@ -139,12 +141,7 @@ def _create_data(height=3, width=3, channels=3, device="cpu"): ...@@ -139,12 +141,7 @@ def _create_data(height=3, width=3, channels=3, device="cpu"):
def _create_data_batch(height=3, width=3, channels=3, num_samples=4, device="cpu"): def _create_data_batch(height=3, width=3, channels=3, num_samples=4, device="cpu"):
# TODO: When all relevant tests are ported to pytest, turn this into a module-level fixture # TODO: When all relevant tests are ported to pytest, turn this into a module-level fixture
batch_tensor = torch.randint( batch_tensor = torch.randint(0, 256, (num_samples, channels, height, width), dtype=torch.uint8, device=device)
0, 256,
(num_samples, channels, height, width),
dtype=torch.uint8,
device=device
)
return batch_tensor return batch_tensor
...@@ -180,8 +177,9 @@ def _assert_equal_tensor_to_pil(tensor, pil_image, msg=None): ...@@ -180,8 +177,9 @@ def _assert_equal_tensor_to_pil(tensor, pil_image, msg=None):
assert_equal(tensor.cpu(), pil_tensor, msg=msg) assert_equal(tensor.cpu(), pil_tensor, msg=msg)
def _assert_approx_equal_tensor_to_pil(tensor, pil_image, tol=1e-5, msg=None, agg_method="mean", def _assert_approx_equal_tensor_to_pil(
allowed_percentage_diff=None): tensor, pil_image, tol=1e-5, msg=None, agg_method="mean", allowed_percentage_diff=None
):
# TODO: we could just merge this into _assert_equal_tensor_to_pil # TODO: we could just merge this into _assert_equal_tensor_to_pil
np_pil_image = np.array(pil_image) np_pil_image = np.array(pil_image)
if np_pil_image.ndim == 2: if np_pil_image.ndim == 2:
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment