Unverified Commit 138fac70 authored by Tolga Cangöz's avatar Tolga Cangöz Committed by GitHub
Browse files

Discourage using deprecated `revision` parameter (#8573)



* Discourage using `revision`

* `make style && make quality`

* Refactor code to use 'variant' instead of 'revision'

* `revision="bf16"` -> `variant="bf16"`

---------
Co-authored-by: default avatarSayak Paul <spsayakpaul@gmail.com>
parent 468ae09e
...@@ -48,7 +48,7 @@ from diffusers import DiffusionPipeline, DPMSolverMultistepScheduler ...@@ -48,7 +48,7 @@ from diffusers import DiffusionPipeline, DPMSolverMultistepScheduler
import torch import torch
repo_id = "stabilityai/stable-diffusion-2-base" repo_id = "stabilityai/stable-diffusion-2-base"
pipe = DiffusionPipeline.from_pretrained(repo_id, torch_dtype=torch.float16, revision="fp16") pipe = DiffusionPipeline.from_pretrained(repo_id, torch_dtype=torch.float16, variant="fp16")
pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config)
pipe = pipe.to("cuda") pipe = pipe.to("cuda")
...@@ -72,7 +72,7 @@ init_image = load_image(img_url).resize((512, 512)) ...@@ -72,7 +72,7 @@ init_image = load_image(img_url).resize((512, 512))
mask_image = load_image(mask_url).resize((512, 512)) mask_image = load_image(mask_url).resize((512, 512))
repo_id = "stabilityai/stable-diffusion-2-inpainting" repo_id = "stabilityai/stable-diffusion-2-inpainting"
pipe = DiffusionPipeline.from_pretrained(repo_id, torch_dtype=torch.float16, revision="fp16") pipe = DiffusionPipeline.from_pretrained(repo_id, torch_dtype=torch.float16, variant="fp16")
pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.scheduler = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config)
pipe = pipe.to("cuda") pipe = pipe.to("cuda")
......
...@@ -186,7 +186,7 @@ scheduler, scheduler_state = FlaxDPMSolverMultistepScheduler.from_pretrained( ...@@ -186,7 +186,7 @@ scheduler, scheduler_state = FlaxDPMSolverMultistepScheduler.from_pretrained(
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
"runwayml/stable-diffusion-v1-5", "runwayml/stable-diffusion-v1-5",
scheduler=scheduler, scheduler=scheduler,
revision="bf16", variant="bf16",
dtype=jax.numpy.bfloat16, dtype=jax.numpy.bfloat16,
) )
params["scheduler"] = scheduler_state params["scheduler"] = scheduler_state
......
...@@ -63,7 +63,7 @@ Flax is a functional framework, so models are stateless and parameters are store ...@@ -63,7 +63,7 @@ Flax is a functional framework, so models are stateless and parameters are store
dtype = jnp.bfloat16 dtype = jnp.bfloat16
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
"CompVis/stable-diffusion-v1-4", "CompVis/stable-diffusion-v1-4",
revision="bf16", variant="bf16",
dtype=dtype, dtype=dtype,
) )
``` ```
......
...@@ -296,7 +296,7 @@ scheduler, scheduler_state = FlaxDPMSolverMultistepScheduler.from_pretrained( ...@@ -296,7 +296,7 @@ scheduler, scheduler_state = FlaxDPMSolverMultistepScheduler.from_pretrained(
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
model_id, model_id,
scheduler=scheduler, scheduler=scheduler,
revision="bf16", variant="bf16",
dtype=jax.numpy.bfloat16, dtype=jax.numpy.bfloat16,
) )
params["scheduler"] = scheduler_state params["scheduler"] = scheduler_state
......
...@@ -83,7 +83,7 @@ Flax는 함수형 프레임워크이므로 모델은 무상태(stateless)형이 ...@@ -83,7 +83,7 @@ Flax는 함수형 프레임워크이므로 모델은 무상태(stateless)형이
```python ```python
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
"CompVis/stable-diffusion-v1-4", "CompVis/stable-diffusion-v1-4",
revision="bf16", variant="bf16",
dtype=dtype, dtype=dtype,
) )
``` ```
......
...@@ -421,7 +421,7 @@ import torch ...@@ -421,7 +421,7 @@ import torch
pipe = DiffusionPipeline.from_pretrained( pipe = DiffusionPipeline.from_pretrained(
"CompVis/stable-diffusion-v1-4", "CompVis/stable-diffusion-v1-4",
revision='fp16', variant='fp16',
torch_dtype=torch.float16, torch_dtype=torch.float16,
safety_checker=None, # Very important for videos...lots of false positives while interpolating safety_checker=None, # Very important for videos...lots of false positives while interpolating
custom_pipeline="interpolate_stable_diffusion", custom_pipeline="interpolate_stable_diffusion",
...@@ -462,7 +462,7 @@ def download_image(url): ...@@ -462,7 +462,7 @@ def download_image(url):
response = requests.get(url) response = requests.get(url)
return PIL.Image.open(BytesIO(response.content)).convert("RGB") return PIL.Image.open(BytesIO(response.content)).convert("RGB")
pipe = DiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", custom_pipeline="stable_diffusion_mega", torch_dtype=torch.float16, revision="fp16") pipe = DiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", custom_pipeline="stable_diffusion_mega", torch_dtype=torch.float16, variant="fp16")
pipe.to("cuda") pipe.to("cuda")
pipe.enable_attention_slicing() pipe.enable_attention_slicing()
...@@ -1506,12 +1506,12 @@ scheduler = DDIMScheduler.from_pretrained("stabilityai/stable-diffusion-2-1", ...@@ -1506,12 +1506,12 @@ scheduler = DDIMScheduler.from_pretrained("stabilityai/stable-diffusion-2-1",
pipe = StableDiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-2-1", pipe = StableDiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-2-1",
custom_pipeline="stable_diffusion_tensorrt_txt2img", custom_pipeline="stable_diffusion_tensorrt_txt2img",
revision='fp16', variant='fp16',
torch_dtype=torch.float16, torch_dtype=torch.float16,
scheduler=scheduler,) scheduler=scheduler,)
# re-use cached folder to save ONNX models and TensorRT Engines # re-use cached folder to save ONNX models and TensorRT Engines
pipe.set_cached_folder("stabilityai/stable-diffusion-2-1", revision='fp16',) pipe.set_cached_folder("stabilityai/stable-diffusion-2-1", variant='fp16',)
pipe = pipe.to("cuda") pipe = pipe.to("cuda")
...@@ -1568,7 +1568,7 @@ text_encoder = CLIPTextModel.from_pretrained( ...@@ -1568,7 +1568,7 @@ text_encoder = CLIPTextModel.from_pretrained(
pipeline = DiffusionPipeline.from_pretrained( pipeline = DiffusionPipeline.from_pretrained(
pretrained_model_name_or_path="CompVis/stable-diffusion-v1-4", pretrained_model_name_or_path="CompVis/stable-diffusion-v1-4",
custom_pipeline="edict_pipeline", custom_pipeline="edict_pipeline",
revision="fp16", variant="fp16",
scheduler=scheduler, scheduler=scheduler,
text_encoder=text_encoder, text_encoder=text_encoder,
leapfrog_steps=True, leapfrog_steps=True,
...@@ -1660,12 +1660,12 @@ scheduler = DDIMScheduler.from_pretrained("stabilityai/stable-diffusion-2-1", ...@@ -1660,12 +1660,12 @@ scheduler = DDIMScheduler.from_pretrained("stabilityai/stable-diffusion-2-1",
pipe = StableDiffusionImg2ImgPipeline.from_pretrained("stabilityai/stable-diffusion-2-1", pipe = StableDiffusionImg2ImgPipeline.from_pretrained("stabilityai/stable-diffusion-2-1",
custom_pipeline="stable_diffusion_tensorrt_img2img", custom_pipeline="stable_diffusion_tensorrt_img2img",
revision='fp16', variant='fp16',
torch_dtype=torch.float16, torch_dtype=torch.float16,
scheduler=scheduler,) scheduler=scheduler,)
# re-use cached folder to save ONNX models and TensorRT Engines # re-use cached folder to save ONNX models and TensorRT Engines
pipe.set_cached_folder("stabilityai/stable-diffusion-2-1", revision='fp16',) pipe.set_cached_folder("stabilityai/stable-diffusion-2-1", variant='fp16',)
pipe = pipe.to("cuda") pipe = pipe.to("cuda")
...@@ -2251,13 +2251,13 @@ scheduler = PNDMScheduler.from_pretrained("stabilityai/stable-diffusion-2-inpain ...@@ -2251,13 +2251,13 @@ scheduler = PNDMScheduler.from_pretrained("stabilityai/stable-diffusion-2-inpain
pipe = StableDiffusionInpaintPipeline.from_pretrained("stabilityai/stable-diffusion-2-inpainting", pipe = StableDiffusionInpaintPipeline.from_pretrained("stabilityai/stable-diffusion-2-inpainting",
custom_pipeline="stable_diffusion_tensorrt_inpaint", custom_pipeline="stable_diffusion_tensorrt_inpaint",
revision='fp16', variant='fp16',
torch_dtype=torch.float16, torch_dtype=torch.float16,
scheduler=scheduler, scheduler=scheduler,
) )
# re-use cached folder to save ONNX models and TensorRT Engines # re-use cached folder to save ONNX models and TensorRT Engines
pipe.set_cached_folder("stabilityai/stable-diffusion-2-inpainting", revision='fp16',) pipe.set_cached_folder("stabilityai/stable-diffusion-2-inpainting", variant='fp16',)
pipe = pipe.to("cuda") pipe = pipe.to("cuda")
......
...@@ -282,7 +282,7 @@ class StableDiffusionTiledUpscalePipeline(StableDiffusionUpscalePipeline): ...@@ -282,7 +282,7 @@ class StableDiffusionTiledUpscalePipeline(StableDiffusionUpscalePipeline):
def main(): def main():
# Run a demo # Run a demo
model_id = "stabilityai/stable-diffusion-x4-upscaler" model_id = "stabilityai/stable-diffusion-x4-upscaler"
pipe = StableDiffusionTiledUpscalePipeline.from_pretrained(model_id, revision="fp16", torch_dtype=torch.float16) pipe = StableDiffusionTiledUpscalePipeline.from_pretrained(model_id, variant="fp16", torch_dtype=torch.float16)
pipe = pipe.to("cuda") pipe = pipe.to("cuda")
image = Image.open("../../docs/source/imgs/diffusers_library.jpg") image = Image.open("../../docs/source/imgs/diffusers_library.jpg")
......
...@@ -296,7 +296,7 @@ class FlaxDiffusionPipeline(ConfigMixin, PushToHubMixin): ...@@ -296,7 +296,7 @@ class FlaxDiffusionPipeline(ConfigMixin, PushToHubMixin):
>>> # see more in [the documentation](https://huggingface.co/docs/hub/security-tokens) >>> # see more in [the documentation](https://huggingface.co/docs/hub/security-tokens)
>>> pipeline, params = FlaxDiffusionPipeline.from_pretrained( >>> pipeline, params = FlaxDiffusionPipeline.from_pretrained(
... "runwayml/stable-diffusion-v1-5", ... "runwayml/stable-diffusion-v1-5",
... revision="bf16", ... variant="bf16",
... dtype=jnp.bfloat16, ... dtype=jnp.bfloat16,
... ) ... )
...@@ -310,7 +310,7 @@ class FlaxDiffusionPipeline(ConfigMixin, PushToHubMixin): ...@@ -310,7 +310,7 @@ class FlaxDiffusionPipeline(ConfigMixin, PushToHubMixin):
... ) ... )
>>> dpm_pipe, dpm_params = FlaxStableDiffusionPipeline.from_pretrained( >>> dpm_pipe, dpm_params = FlaxStableDiffusionPipeline.from_pretrained(
... model_id, revision="bf16", dtype=jnp.bfloat16, scheduler=dpmpp ... model_id, variant="bf16", dtype=jnp.bfloat16, scheduler=dpmpp
... ) ... )
>>> dpm_params["scheduler"] = dpmpp_state >>> dpm_params["scheduler"] = dpmpp_state
``` ```
...@@ -564,7 +564,7 @@ class FlaxDiffusionPipeline(ConfigMixin, PushToHubMixin): ...@@ -564,7 +564,7 @@ class FlaxDiffusionPipeline(ConfigMixin, PushToHubMixin):
... ) ... )
>>> text2img = FlaxStableDiffusionPipeline.from_pretrained( >>> text2img = FlaxStableDiffusionPipeline.from_pretrained(
... "runwayml/stable-diffusion-v1-5", revision="bf16", dtype=jnp.bfloat16 ... "runwayml/stable-diffusion-v1-5", variant="bf16", dtype=jnp.bfloat16
... ) ... )
>>> img2img = FlaxStableDiffusionImg2ImgPipeline(**text2img.components) >>> img2img = FlaxStableDiffusionImg2ImgPipeline(**text2img.components)
``` ```
......
...@@ -55,7 +55,7 @@ EXAMPLE_DOC_STRING = """ ...@@ -55,7 +55,7 @@ EXAMPLE_DOC_STRING = """
>>> from diffusers import FlaxStableDiffusionPipeline >>> from diffusers import FlaxStableDiffusionPipeline
>>> pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( >>> pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
... "runwayml/stable-diffusion-v1-5", revision="bf16", dtype=jax.numpy.bfloat16 ... "runwayml/stable-diffusion-v1-5", variant="bf16", dtype=jax.numpy.bfloat16
... ) ... )
>>> prompt = "a photo of an astronaut riding a horse on mars" >>> prompt = "a photo of an astronaut riding a horse on mars"
......
...@@ -612,7 +612,7 @@ class StableDiffusionUpscalePipeline( ...@@ -612,7 +612,7 @@ class StableDiffusionUpscalePipeline(
>>> # load model and scheduler >>> # load model and scheduler
>>> model_id = "stabilityai/stable-diffusion-x4-upscaler" >>> model_id = "stabilityai/stable-diffusion-x4-upscaler"
>>> pipeline = StableDiffusionUpscalePipeline.from_pretrained( >>> pipeline = StableDiffusionUpscalePipeline.from_pretrained(
... model_id, revision="fp16", torch_dtype=torch.float16 ... model_id, variant="fp16", torch_dtype=torch.float16
... ) ... )
>>> pipeline = pipeline.to("cuda") >>> pipeline = pipeline.to("cuda")
......
...@@ -1013,7 +1013,7 @@ class StableDiffusionPipelineSlowTests(unittest.TestCase): ...@@ -1013,7 +1013,7 @@ class StableDiffusionPipelineSlowTests(unittest.TestCase):
torch.cuda.reset_peak_memory_stats() torch.cuda.reset_peak_memory_stats()
model_id = "CompVis/stable-diffusion-v1-4" model_id = "CompVis/stable-diffusion-v1-4"
pipe = StableDiffusionPipeline.from_pretrained( pipe = StableDiffusionPipeline.from_pretrained(
model_id, revision="fp16", torch_dtype=torch.float16, safety_checker=None model_id, variant="fp16", torch_dtype=torch.float16, safety_checker=None
) )
pipe.set_progress_bar_config(disable=None) pipe.set_progress_bar_config(disable=None)
pipe.enable_attention_slicing() pipe.enable_attention_slicing()
......
...@@ -39,7 +39,7 @@ class FlaxStableDiffusion2PipelineIntegrationTests(unittest.TestCase): ...@@ -39,7 +39,7 @@ class FlaxStableDiffusion2PipelineIntegrationTests(unittest.TestCase):
def test_stable_diffusion_flax(self): def test_stable_diffusion_flax(self):
sd_pipe, params = FlaxStableDiffusionPipeline.from_pretrained( sd_pipe, params = FlaxStableDiffusionPipeline.from_pretrained(
"stabilityai/stable-diffusion-2", "stabilityai/stable-diffusion-2",
revision="bf16", variant="bf16",
dtype=jnp.bfloat16, dtype=jnp.bfloat16,
) )
...@@ -80,7 +80,7 @@ class FlaxStableDiffusion2PipelineNightlyTests(unittest.TestCase): ...@@ -80,7 +80,7 @@ class FlaxStableDiffusion2PipelineNightlyTests(unittest.TestCase):
sd_pipe, params = FlaxStableDiffusionPipeline.from_pretrained( sd_pipe, params = FlaxStableDiffusionPipeline.from_pretrained(
model_id, model_id,
scheduler=scheduler, scheduler=scheduler,
revision="bf16", variant="bf16",
dtype=jnp.bfloat16, dtype=jnp.bfloat16,
) )
params["scheduler"] = scheduler_params params["scheduler"] = scheduler_params
......
...@@ -115,7 +115,7 @@ class FlaxPipelineTests(unittest.TestCase): ...@@ -115,7 +115,7 @@ class FlaxPipelineTests(unittest.TestCase):
def test_stable_diffusion_v1_4_bfloat_16(self): def test_stable_diffusion_v1_4_bfloat_16(self):
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
"CompVis/stable-diffusion-v1-4", revision="bf16", dtype=jnp.bfloat16, safety_checker=None "CompVis/stable-diffusion-v1-4", variant="bf16", dtype=jnp.bfloat16, safety_checker=None
) )
prompt = ( prompt = (
...@@ -144,7 +144,7 @@ class FlaxPipelineTests(unittest.TestCase): ...@@ -144,7 +144,7 @@ class FlaxPipelineTests(unittest.TestCase):
def test_stable_diffusion_v1_4_bfloat_16_with_safety(self): def test_stable_diffusion_v1_4_bfloat_16_with_safety(self):
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
"CompVis/stable-diffusion-v1-4", revision="bf16", dtype=jnp.bfloat16 "CompVis/stable-diffusion-v1-4", variant="bf16", dtype=jnp.bfloat16
) )
prompt = ( prompt = (
...@@ -182,7 +182,7 @@ class FlaxPipelineTests(unittest.TestCase): ...@@ -182,7 +182,7 @@ class FlaxPipelineTests(unittest.TestCase):
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
"CompVis/stable-diffusion-v1-4", "CompVis/stable-diffusion-v1-4",
revision="bf16", variant="bf16",
dtype=jnp.bfloat16, dtype=jnp.bfloat16,
scheduler=scheduler, scheduler=scheduler,
safety_checker=None, safety_checker=None,
...@@ -227,7 +227,7 @@ class FlaxPipelineTests(unittest.TestCase): ...@@ -227,7 +227,7 @@ class FlaxPipelineTests(unittest.TestCase):
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
"CompVis/stable-diffusion-v1-4", "CompVis/stable-diffusion-v1-4",
revision="bf16", variant="bf16",
dtype=jnp.bfloat16, dtype=jnp.bfloat16,
safety_checker=None, safety_checker=None,
) )
...@@ -242,7 +242,7 @@ class FlaxPipelineTests(unittest.TestCase): ...@@ -242,7 +242,7 @@ class FlaxPipelineTests(unittest.TestCase):
# With memory efficient attention # With memory efficient attention
pipeline, params = FlaxStableDiffusionPipeline.from_pretrained( pipeline, params = FlaxStableDiffusionPipeline.from_pretrained(
"CompVis/stable-diffusion-v1-4", "CompVis/stable-diffusion-v1-4",
revision="bf16", variant="bf16",
dtype=jnp.bfloat16, dtype=jnp.bfloat16,
safety_checker=None, safety_checker=None,
use_memory_efficient_attention=True, use_memory_efficient_attention=True,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment