using_safetensors.md 4.69 KB
Newer Older
Steven Liu's avatar
Steven Liu committed
1
# Load safetensors
2

3
4
[[open-in-colab]]

Steven Liu's avatar
Steven Liu committed
5
[safetensors](https://github.com/huggingface/safetensors) is a safe and fast file format for storing and loading tensors. Typically, PyTorch model weights are saved or *pickled* into a `.bin` file with Python's [`pickle`](https://docs.python.org/3/library/pickle.html) utility. However, `pickle` is not secure and pickled files may contain malicious code that can be executed. safetensors is a secure alternative to `pickle`, making it ideal for sharing model weights.
6

Steven Liu's avatar
Steven Liu committed
7
This guide will show you how you load `.safetensor` files, and how to convert Stable Diffusion model weights stored in other formats to `.safetensor`. Before you start, make sure you have safetensors installed:
8

9
10
11
```py
# uncomment to install the necessary libraries in Colab
#!pip install safetensors
12
13
```

Steven Liu's avatar
Steven Liu committed
14
If you look at the [`runwayml/stable-diffusion-v1-5`](https://huggingface.co/runwayml/stable-diffusion-v1-5/tree/main) repository, you'll see weights inside the `text_encoder`, `unet` and `vae` subfolders are stored in the `.safetensors` format. By default, 🤗 Diffusers automatically loads these `.safetensors` files from their subfolders if they're available in the model repository.
15

Steven Liu's avatar
Steven Liu committed
16
For more explicit control, you can optionally set `use_safetensors=True` (if `safetensors` is not installed, you'll get an error message asking you to install it):
17

Steven Liu's avatar
Steven Liu committed
18
19
```py
from diffusers import DiffusionPipeline
20

Steven Liu's avatar
Steven Liu committed
21
pipeline = DiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", use_safetensors=True)
22
23
```

Patrick von Platen's avatar
Patrick von Platen committed
24
However, model weights are not necessarily stored in separate subfolders like in the example above. Sometimes, all the weights are stored in a single `.safetensors` file. In this case, if the weights are Stable Diffusion weights, you can load the file directly with the [`~diffusers.loaders.FromSingleFileMixin.from_single_file`] method:
25

Steven Liu's avatar
Steven Liu committed
26
27
```py
from diffusers import StableDiffusionPipeline
28

Patrick von Platen's avatar
Patrick von Platen committed
29
pipeline = StableDiffusionPipeline.from_single_file(
Steven Liu's avatar
Steven Liu committed
30
31
32
    "https://huggingface.co/WarriorMama777/OrangeMixs/blob/main/Models/AbyssOrangeMix/AbyssOrangeMix.safetensors"
)
```
33

Steven Liu's avatar
Steven Liu committed
34
## Convert to safetensors
35

Steven Liu's avatar
Steven Liu committed
36
Not all weights on the Hub are available in the `.safetensors` format, and you may encounter weights stored as `.bin`. In this case, use the [Convert Space](https://huggingface.co/spaces/diffusers/convert) to convert the weights to `.safetensors`. The Convert Space downloads the pickled weights, converts them, and opens a Pull Request to upload the newly converted `.safetensors` file on the Hub. This way, if there is any malicious code contained in the pickled files, they're uploaded to the Hub - which has a [security scanner](https://huggingface.co/docs/hub/security-pickle#hubs-security-scanner) to detect unsafe files and suspicious pickle imports - instead of your computer.
37

Steven Liu's avatar
Steven Liu committed
38
You can use the model with the new `.safetensors` weights by specifying the reference to the Pull Request in the `revision` parameter (you can also test it in this [Check PR](https://huggingface.co/spaces/diffusers/check_pr) Space on the Hub), for example `refs/pr/22`:
39

Steven Liu's avatar
Steven Liu committed
40
41
```py
from diffusers import DiffusionPipeline
42

43
44
45
pipeline = DiffusionPipeline.from_pretrained(
    "stabilityai/stable-diffusion-2-1", revision="refs/pr/22", use_safetensors=True
)
46
47
```

Steven Liu's avatar
Steven Liu committed
48
## Why use safetensors?
49

Steven Liu's avatar
Steven Liu committed
50
There are several reasons for using safetensors:
51

Steven Liu's avatar
Steven Liu committed
52
53
- Safety is the number one reason for using safetensors. As open-source and model distribution grows, it is important to be able to trust the model weights you downloaded don't contain any malicious code. The current size of the header in safetensors prevents parsing extremely large JSON files.
- Loading speed between switching models is another reason to use safetensors, which performs zero-copy of the tensors. It is especially fast compared to `pickle` if you're loading the weights to CPU (the default case), and just as fast if not faster when directly loading the weights to GPU. You'll only notice the performance difference if the model is already loaded, and not if you're downloading the weights or loading the model for the first time.
54

Steven Liu's avatar
Steven Liu committed
55
	The time it takes to load the entire pipeline:
56

Steven Liu's avatar
Steven Liu committed
57
	```py
Patrick von Platen's avatar
Patrick von Platen committed
58
 from diffusers import StableDiffusionPipeline
59

60
 pipeline = StableDiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-2-1", use_safetensors=True)
Patrick von Platen's avatar
Patrick von Platen committed
61
62
 "Loaded in safetensors 0:00:02.033658"
 "Loaded in PyTorch 0:00:02.663379"
Steven Liu's avatar
Steven Liu committed
63
	```
64

Steven Liu's avatar
Steven Liu committed
65
	But the actual time it takes to load 500MB of the model weights is only:
66

Steven Liu's avatar
Steven Liu committed
67
68
69
70
	```bash
	safetensors: 3.4873ms
	PyTorch: 172.7537ms
	```
71

Steven Liu's avatar
Steven Liu committed
72
- Lazy loading is also supported in safetensors, which is useful in distributed settings to only load some of the tensors. This format allowed the [BLOOM](https://huggingface.co/bigscience/bloom) model to be loaded in 45 seconds on 8 GPUs instead of 10 minutes with regular PyTorch weights.