"git@developer.sourcefind.cn:chenpangpang/transformers.git" did not exist on "4eec5d0cf67116e98770c305640b5710571da4f6"
Unverified Commit 087fd5f3 authored by Younes Belkada's avatar Younes Belkada Committed by GitHub
Browse files

[`ImageProcessor`] Refactor default `mean` & `std` to `OPENAI_CLIP_MEAN` &...

[`ImageProcessor`] Refactor default `mean` & `std` to `OPENAI_CLIP_MEAN` & `OPENAI_CLIP_STD` (#21425)

* fix default value

* add the fix on other models
parent 005b5157
...@@ -35,6 +35,8 @@ from .utils.constants import ( # noqa: F401 ...@@ -35,6 +35,8 @@ from .utils.constants import ( # noqa: F401
IMAGENET_DEFAULT_STD, IMAGENET_DEFAULT_STD,
IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_MEAN,
IMAGENET_STANDARD_STD, IMAGENET_STANDARD_STD,
OPENAI_CLIP_MEAN,
OPENAI_CLIP_STD,
) )
......
...@@ -31,6 +31,8 @@ from ...image_transforms import ( ...@@ -31,6 +31,8 @@ from ...image_transforms import (
to_channel_dimension_format, to_channel_dimension_format,
) )
from ...image_utils import ( from ...image_utils import (
OPENAI_CLIP_MEAN,
OPENAI_CLIP_STD,
ChannelDimension, ChannelDimension,
ImageInput, ImageInput,
PILImageResampling, PILImageResampling,
...@@ -118,8 +120,8 @@ class BitImageProcessor(BaseImageProcessor): ...@@ -118,8 +120,8 @@ class BitImageProcessor(BaseImageProcessor):
self.do_rescale = do_rescale self.do_rescale = do_rescale
self.rescale_factor = rescale_factor self.rescale_factor = rescale_factor
self.do_normalize = do_normalize self.do_normalize = do_normalize
self.image_mean = image_mean if image_mean is not None else [0.48145466, 0.4578275, 0.40821073] self.image_mean = image_mean if image_mean is not None else OPENAI_CLIP_MEAN
self.image_std = image_std if image_std is not None else [0.26862954, 0.26130258, 0.27577711] self.image_std = image_std if image_std is not None else OPENAI_CLIP_STD
self.do_convert_rgb = do_convert_rgb self.do_convert_rgb = do_convert_rgb
def resize( def resize(
......
...@@ -24,8 +24,8 @@ from transformers.utils.generic import TensorType ...@@ -24,8 +24,8 @@ from transformers.utils.generic import TensorType
from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict
from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format
from ...image_utils import ( from ...image_utils import (
IMAGENET_STANDARD_MEAN, OPENAI_CLIP_MEAN,
IMAGENET_STANDARD_STD, OPENAI_CLIP_STD,
ChannelDimension, ChannelDimension,
ImageInput, ImageInput,
PILImageResampling, PILImageResampling,
...@@ -103,8 +103,8 @@ class BlipImageProcessor(BaseImageProcessor): ...@@ -103,8 +103,8 @@ class BlipImageProcessor(BaseImageProcessor):
self.do_rescale = do_rescale self.do_rescale = do_rescale
self.rescale_factor = rescale_factor self.rescale_factor = rescale_factor
self.do_normalize = do_normalize self.do_normalize = do_normalize
self.image_mean = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN self.image_mean = image_mean if image_mean is not None else OPENAI_CLIP_MEAN
self.image_std = image_std if image_std is not None else IMAGENET_STANDARD_STD self.image_std = image_std if image_std is not None else OPENAI_CLIP_STD
self.do_convert_rgb = do_convert_rgb self.do_convert_rgb = do_convert_rgb
def resize( def resize(
......
...@@ -25,6 +25,8 @@ from transformers.utils.generic import TensorType ...@@ -25,6 +25,8 @@ from transformers.utils.generic import TensorType
from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict
from ...image_transforms import PaddingMode, center_crop, normalize, pad, rescale, resize, to_channel_dimension_format from ...image_transforms import PaddingMode, center_crop, normalize, pad, rescale, resize, to_channel_dimension_format
from ...image_utils import ( from ...image_utils import (
OPENAI_CLIP_MEAN,
OPENAI_CLIP_STD,
ChannelDimension, ChannelDimension,
ImageInput, ImageInput,
PILImageResampling, PILImageResampling,
...@@ -186,8 +188,8 @@ class BridgeTowerImageProcessor(BaseImageProcessor): ...@@ -186,8 +188,8 @@ class BridgeTowerImageProcessor(BaseImageProcessor):
self.do_rescale = do_rescale self.do_rescale = do_rescale
self.rescale_factor = rescale_factor self.rescale_factor = rescale_factor
self.do_normalize = do_normalize self.do_normalize = do_normalize
self.image_mean = image_mean if image_mean is not None else [0.48145466, 0.4578275, 0.40821073] self.image_mean = image_mean if image_mean is not None else OPENAI_CLIP_MEAN
self.image_std = image_std if image_std is not None else [0.26862954, 0.26130258, 0.27577711] self.image_std = image_std if image_std is not None else OPENAI_CLIP_STD
self.do_pad = do_pad self.do_pad = do_pad
self.do_center_crop = do_center_crop self.do_center_crop = do_center_crop
......
...@@ -31,6 +31,8 @@ from ...image_transforms import ( ...@@ -31,6 +31,8 @@ from ...image_transforms import (
to_channel_dimension_format, to_channel_dimension_format,
) )
from ...image_utils import ( from ...image_utils import (
OPENAI_CLIP_MEAN,
OPENAI_CLIP_STD,
ChannelDimension, ChannelDimension,
ImageInput, ImageInput,
PILImageResampling, PILImageResampling,
...@@ -118,8 +120,8 @@ class ChineseCLIPImageProcessor(BaseImageProcessor): ...@@ -118,8 +120,8 @@ class ChineseCLIPImageProcessor(BaseImageProcessor):
self.do_rescale = do_rescale self.do_rescale = do_rescale
self.rescale_factor = rescale_factor self.rescale_factor = rescale_factor
self.do_normalize = do_normalize self.do_normalize = do_normalize
self.image_mean = image_mean if image_mean is not None else [0.48145466, 0.4578275, 0.40821073] self.image_mean = image_mean if image_mean is not None else OPENAI_CLIP_MEAN
self.image_std = image_std if image_std is not None else [0.26862954, 0.26130258, 0.27577711] self.image_std = image_std if image_std is not None else OPENAI_CLIP_STD
self.do_convert_rgb = do_convert_rgb self.do_convert_rgb = do_convert_rgb
def resize( def resize(
......
...@@ -31,6 +31,8 @@ from ...image_transforms import ( ...@@ -31,6 +31,8 @@ from ...image_transforms import (
to_channel_dimension_format, to_channel_dimension_format,
) )
from ...image_utils import ( from ...image_utils import (
OPENAI_CLIP_MEAN,
OPENAI_CLIP_STD,
ChannelDimension, ChannelDimension,
ImageInput, ImageInput,
PILImageResampling, PILImageResampling,
...@@ -118,8 +120,8 @@ class CLIPImageProcessor(BaseImageProcessor): ...@@ -118,8 +120,8 @@ class CLIPImageProcessor(BaseImageProcessor):
self.do_rescale = do_rescale self.do_rescale = do_rescale
self.rescale_factor = rescale_factor self.rescale_factor = rescale_factor
self.do_normalize = do_normalize self.do_normalize = do_normalize
self.image_mean = image_mean if image_mean is not None else [0.48145466, 0.4578275, 0.40821073] self.image_mean = image_mean if image_mean is not None else OPENAI_CLIP_MEAN
self.image_std = image_std if image_std is not None else [0.26862954, 0.26130258, 0.27577711] self.image_std = image_std if image_std is not None else OPENAI_CLIP_STD
self.do_convert_rgb = do_convert_rgb self.do_convert_rgb = do_convert_rgb
def resize( def resize(
......
...@@ -27,6 +27,8 @@ from transformers.utils.generic import TensorType ...@@ -27,6 +27,8 @@ from transformers.utils.generic import TensorType
from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict
from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format
from ...image_utils import ( from ...image_utils import (
OPENAI_CLIP_MEAN,
OPENAI_CLIP_STD,
ChannelDimension, ChannelDimension,
ImageInput, ImageInput,
PILImageResampling, PILImageResampling,
...@@ -45,8 +47,8 @@ logger = logging.get_logger(__name__) ...@@ -45,8 +47,8 @@ logger = logging.get_logger(__name__)
# These values are taken from CLIP # These values are taken from CLIP
FLAVA_IMAGE_MEAN = [0.48145466, 0.4578275, 0.40821073] FLAVA_IMAGE_MEAN = OPENAI_CLIP_MEAN
FLAVA_IMAGE_STD = [0.26862954, 0.26130258, 0.27577711] FLAVA_IMAGE_STD = OPENAI_CLIP_STD
FLAVA_CODEBOOK_MEAN = [0.0, 0.0, 0.0] FLAVA_CODEBOOK_MEAN = [0.0, 0.0, 0.0]
FLAVA_CODEBOOK_STD = [1.0, 1.0, 1.0] FLAVA_CODEBOOK_STD = [1.0, 1.0, 1.0]
LOGIT_LAPLACE_EPS: float = 0.1 LOGIT_LAPLACE_EPS: float = 0.1
......
...@@ -30,6 +30,8 @@ from transformers.image_transforms import ( ...@@ -30,6 +30,8 @@ from transformers.image_transforms import (
to_numpy_array, to_numpy_array,
) )
from transformers.image_utils import ( from transformers.image_utils import (
OPENAI_CLIP_MEAN,
OPENAI_CLIP_STD,
ChannelDimension, ChannelDimension,
ImageInput, ImageInput,
PILImageResampling, PILImageResampling,
...@@ -160,8 +162,8 @@ class OwlViTImageProcessor(BaseImageProcessor): ...@@ -160,8 +162,8 @@ class OwlViTImageProcessor(BaseImageProcessor):
self.do_rescale = do_rescale self.do_rescale = do_rescale
self.rescale_factor = rescale_factor self.rescale_factor = rescale_factor
self.do_normalize = do_normalize self.do_normalize = do_normalize
self.image_mean = image_mean if image_mean is not None else [0.48145466, 0.4578275, 0.40821073] self.image_mean = image_mean if image_mean is not None else OPENAI_CLIP_MEAN
self.image_std = image_std if image_std is not None else [0.26862954, 0.26130258, 0.27577711] self.image_std = image_std if image_std is not None else OPENAI_CLIP_STD
def resize( def resize(
self, self,
......
...@@ -31,6 +31,8 @@ from ...image_transforms import ( ...@@ -31,6 +31,8 @@ from ...image_transforms import (
to_channel_dimension_format, to_channel_dimension_format,
) )
from ...image_utils import ( from ...image_utils import (
OPENAI_CLIP_MEAN,
OPENAI_CLIP_STD,
ChannelDimension, ChannelDimension,
ImageInput, ImageInput,
PILImageResampling, PILImageResampling,
...@@ -118,8 +120,8 @@ class ViTHybridImageProcessor(BaseImageProcessor): ...@@ -118,8 +120,8 @@ class ViTHybridImageProcessor(BaseImageProcessor):
self.do_rescale = do_rescale self.do_rescale = do_rescale
self.rescale_factor = rescale_factor self.rescale_factor = rescale_factor
self.do_normalize = do_normalize self.do_normalize = do_normalize
self.image_mean = image_mean if image_mean is not None else [0.48145466, 0.4578275, 0.40821073] self.image_mean = image_mean if image_mean is not None else OPENAI_CLIP_MEAN
self.image_std = image_std if image_std is not None else [0.26862954, 0.26130258, 0.27577711] self.image_std = image_std if image_std is not None else OPENAI_CLIP_STD
self.do_convert_rgb = do_convert_rgb self.do_convert_rgb = do_convert_rgb
def resize( def resize(
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment