installation.mdx 17 KB
Newer Older
1
# Installation Guide
Titus's avatar
Titus committed
2

3
Welcome to the installation guide for the `bitsandbytes` library! This document provides step-by-step instructions to install `bitsandbytes` across various platforms and hardware configurations. The library primarily supports CUDA-based GPUs, but the team is actively working on enabling support for additional backends like AMD ROCm, Intel, and Apple Silicon.
jiqing-feng's avatar
jiqing-feng committed
4

5
6
> [!TIP]
> For a high-level overview of backend support and compatibility, see the [Multi-backend Support](#multi-backend) section.
Younes Belkada's avatar
Younes Belkada committed
7

8
## Table of Contents
9

10
11
12
13
14
15
16
17
18
- [CUDA](#cuda)
  - [Installation via PyPI](#cuda-pip)
  - [Compile from Source](#cuda-compile)
- [Multi-backend Support (Alpha Release)](#multi-backend)
  - [Supported Backends](#multi-backend-supported-backends)
  - [Pre-requisites](#multi-backend-pre-requisites)
  - [Installation](#multi-backend-pip)
  - [Compile from Source](#multi-backend-compile)
- [PyTorch CUDA Versions](#pytorch-cuda-versions)
19

20
## CUDA[[cuda]]
Younes Belkada's avatar
Younes Belkada committed
21

22
`bitsandbytes` is currently only supported on CUDA GPUs for CUDA versions **11.0 - 12.6**. However, there's an ongoing multi-backend effort under development, which is currently in alpha. If you're interested in providing feedback or testing, check out [the multi-backend section below](#multi-backend).
Titus's avatar
Titus committed
23

24
25
### Supported CUDA Configurations[[cuda-pip]]

26
The latest version of the distributed `bitsandbytes` package is built with the following configurations:
27

28
| **OS**      | **CUDA Toolkit** | **Host Compiler**         |
29
30
|-------------|------------------|----------------------|
| **Linux**   | 11.7 - 12.3      | GCC 11.4             |
31
32
|             | 12.4 - 12.6      | GCC 13.2             |
| **Windows** | 11.7 - 12.6      | MSVC 19.42+ (VS2022) |
33
|             | 12.4+            | GCC 13.2             |
34
| **Windows** | 11.7 - 12.6      | MSVC 19.38+ (VS2022) |
35

36
For CUDA systems, ensure your hardware meets the following requirements:
37

38
39
40
41
42
| **Feature**                     | **Minimum Hardware Requirement**                              |
|---------------------------------|---------------------------------------------------------------|
| LLM.int8()                      | NVIDIA Turing (RTX 20 series, T4) or newer GPUs               |
| 8-bit optimizers/quantization   | NVIDIA Maxwell (GTX 900 series, TITAN X, M40) or newer GPUs * |
| NF4/FP4 quantization            | NVIDIA Maxwell (GTX 900 series, TITAN X, M40) or newer GPUs * |
Titus's avatar
Titus committed
43

Steven Liu's avatar
Steven Liu committed
44
> [!WARNING]
45
46
47
> `bitsandbytes >= 0.45.0` no longer supports Kepler GPUs.
>
> Support for Maxwell GPUs is deprecated and will be removed in a future release. For the best results, a Turing generation device or newer is recommended.
Younes Belkada's avatar
Younes Belkada committed
48
49
50
51
52

```bash
pip install bitsandbytes
```

53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
### `pip install` pre-built wheel from latest `main` commit

If you would like to use new feature even before they are officially released and help us test them, feel free to install the wheel directly from our CI (*the wheel links will remain stable!*):

<hfoptions id="OS">
<hfoption id="Linux">

```
# Note, if you don't want to reinstall BNBs dependencies, append the `--no-deps` flag!
pip install --force-reinstall 'https://github.com/bitsandbytes-foundation/bitsandbytes/releases/download/continuous-release_main/bitsandbytes-0.44.2.dev0-py3-none-manylinux_2_24_x86_64.whl'
```

</hfoption>
<hfoption id="Windows">

```
# Note, if you don't want to reinstall BNBs dependencies, append the `--no-deps` flag!
pip install --force-reinstall 'https://github.com/bitsandbytes-foundation/bitsandbytes/releases/download/continuous-release_multi-backend-refactor/bitsandbytes-0.44.1.dev0-py3-none-macosx_13_1_arm64.whl'
```
</hfoption>
</hfoptions>

### Compile from source[[cuda-compile]]

> [!TIP]
> Don't hesitate to compile from source! The process is pretty straight forward and resilient. This might be needed for older CUDA versions or other less common configurations, which we don't support out of the box due to package size.
79

80
For Linux and Windows systems, compiling from source allows you to customize the build configurations. See below for detailed platform-specific instructions (see the `CMakeLists.txt` if you want to check the specifics and explore some additional options):
81
82
83
84

<hfoptions id="source">
<hfoption id="Linux">

85
To compile from source, you need CMake >= **3.22.1** and Python >= **3.9** installed. Make sure you have a compiler installed to compile C++ (`gcc`, `make`, headers, etc.).
86
87

For example, to install a compiler and CMake on Ubuntu:
Younes Belkada's avatar
Younes Belkada committed
88

Steven Liu's avatar
Steven Liu committed
89
90
91
92
```bash
apt-get install -y build-essential cmake
```

93
94
95
96
You should also install CUDA Toolkit by following the [NVIDIA CUDA Installation Guide for Linux](https://docs.nvidia.com/cuda/cuda-installation-guide-linux/index.html) guide from NVIDIA. The current expected CUDA Toolkit version is **11.1+** and it is recommended to install **GCC >= 7.3** and required to have at least **GCC >= 6**.

Refer to the following table if you're using another CUDA Toolkit version.

97
98
99
100
101
| CUDA Toolkit |  GCC  |
|--------------|-------|
| >= 11.4.1    | >= 11 |
| >= 12.0      | >= 12 |
| >= 12.4      | >= 13 |
Steven Liu's avatar
Steven Liu committed
102
103

Now to install the bitsandbytes package from source, run the following commands:
104

Younes Belkada's avatar
Younes Belkada committed
105
```bash
106
git clone https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/
107
108
109
pip install -r requirements-dev.txt
cmake -DCOMPUTE_BACKEND=cuda -S .
make
110
pip install -e .   # `-e` for "editable" install, when developing BNB (otherwise leave that out)
Younes Belkada's avatar
Younes Belkada committed
111
```
Steven Liu's avatar
Steven Liu committed
112
113
114

> [!TIP]
> If you have multiple versions of CUDA installed or installed it in a non-standard location, please refer to CMake CUDA documentation for how to configure the CUDA compiler.
Younes Belkada's avatar
Younes Belkada committed
115
116
117
118

</hfoption>
<hfoption id="Windows">

Steven Liu's avatar
Steven Liu committed
119
Windows systems require Visual Studio with C++ support as well as an installation of the CUDA SDK.
120

121
To compile from source, you need CMake >= **3.22.1** and Python >= **3.9** installed. You should also install CUDA Toolkit by following the [CUDA Installation Guide for Windows](https://docs.nvidia.com/cuda/cuda-installation-guide-microsoft-windows/index.html) guide from NVIDIA.
122
123
124
125
126
127

Refer to the following table if you're using another CUDA Toolkit version.

| CUDA Toolkit | MSVC |
|---|---|
| >= 11.6 | 19.30+ (VS2022) |
Younes Belkada's avatar
Younes Belkada committed
128
129

```bash
130
git clone https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/
131
132
cmake -DCOMPUTE_BACKEND=cuda -S .
cmake --build . --config Release
133
pip install -e .   # `-e` for "editable" install, when developing BNB (otherwise leave that out)
Younes Belkada's avatar
Younes Belkada committed
134
135
```

136
Big thanks to [wkpark](https://github.com/wkpark), [Jamezo97](https://github.com/Jamezo97), [rickardp](https://github.com/rickardp), [akx](https://github.com/akx) for their amazing contributions to make bitsandbytes compatible with Windows.
Younes Belkada's avatar
Younes Belkada committed
137

Titus's avatar
Titus committed
138
</hfoption>
Younes Belkada's avatar
Younes Belkada committed
139
</hfoptions>
Steven Liu's avatar
Steven Liu committed
140

141
### PyTorch CUDA versions[[pytorch-cuda-versions]]
Steven Liu's avatar
Steven Liu committed
142
143
144
145
146
147
148
149
150
151
152

Some bitsandbytes features may need a newer CUDA version than the one currently supported by PyTorch binaries from Conda and pip. In this case, you should follow these instructions to load a precompiled bitsandbytes binary.

1. Determine the path of the CUDA version you want to use. Common paths include:

* `/usr/local/cuda`
* `/usr/local/cuda-XX.X` where `XX.X` is the CUDA version number

Then locally install the CUDA version you need with this script from bitsandbytes:

```bash
153
wget https://raw.githubusercontent.com/bitsandbytes-foundation/bitsandbytes/main/install_cuda.sh
Steven Liu's avatar
Steven Liu committed
154
# Syntax cuda_install CUDA_VERSION INSTALL_PREFIX EXPORT_TO_BASH
155
#   CUDA_VERSION in {110, 111, 112, 113, 114, 115, 116, 117, 118, 120, 121, 122, 123, 124, 125, 126}
Steven Liu's avatar
Steven Liu committed
156
157
#   EXPORT_TO_BASH in {0, 1} with 0=False and 1=True

158
# For example, the following installs CUDA 12.6 to ~/local/cuda-12.6 and exports the path to your .bashrc
Steven Liu's avatar
Steven Liu committed
159

160
bash install_cuda.sh 126 ~/local 1
Steven Liu's avatar
Steven Liu committed
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
```

2. Set the environment variables `BNB_CUDA_VERSION` and `LD_LIBRARY_PATH` by manually overriding the CUDA version installed by PyTorch.

> [!TIP]
> It is recommended to add the following lines to the `.bashrc` file to make them permanent.

```bash
export BNB_CUDA_VERSION=<VERSION>
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:<PATH>
```

For example, to use a local install path:

```bash
176
177
export BNB_CUDA_VERSION=126
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/home/YOUR_USERNAME/local/cuda-12.6
Steven Liu's avatar
Steven Liu committed
178
179
180
```

3. Now when you launch bitsandbytes with these environment variables, the PyTorch CUDA version is overridden by the new CUDA version (in this example, version 11.7) and a different bitsandbytes library is loaded.
jiqing-feng's avatar
jiqing-feng committed
181

182
## Multi-backend Support (Alpha Release)[[multi-backend]]
183
184

> [!TIP]
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
> This functionality is currently in preview and not yet production-ready. We very much welcome community feedback, contributions and leadership on topics like Apple Silicon as well as other less common accellerators! For more information, see [this guide on multi-backend support](./non_cuda_backends).

**Link to give us feedback** (bugs, install issues, perf results, requests, etc.)**:**

<hfoptions id="platform">
<hfoption id="ROCm">

[**Multi-backend refactor: Alpha release (AMD ROCm ONLY)**](https://github.com/bitsandbytes-foundation/bitsandbytes/discussions/1339)

</hfoption>
<hfoption id="Intel CPU+GPU">

[**Multi-backend refactor: Alpha release (INTEL ONLY)**](https://github.com/bitsandbytes-foundation/bitsandbytes/discussions/1338)

</hfoption>
<hfoption id="Apple Silicon / Metal (MPS)">
jiqing-feng's avatar
jiqing-feng committed
201

202
[**Github Discussion space on coordinating the kickoff of MPS backend development**](https://github.com/bitsandbytes-foundation/bitsandbytes/discussions/1340)
jiqing-feng's avatar
jiqing-feng committed
203

204
205
</hfoption>
</hfoptions>
206

207
### Supported Backends[[multi-backend-supported-backends]]
208

209
210
211
212
213
214
| **Backend** | **Supported Versions** | **Python versions** | **Architecture Support** | **Status** |
|-------------|------------------------|---------------------------|-------------------------|------------|
| **AMD ROCm** | 6.1+                   | 3.10+                     | minimum CDNA - `gfx90a`, RDNA - `gfx1100` | Alpha      |
| **Apple Silicon (MPS)** | WIP                        | 3.10+                     | M1/M2 chips                    | Planned    |
| **Intel CPU** | v2.4.0+ (`ipex`)         | 3.10+                     | Intel CPU | Alpha |
| **Intel GPU** | v2.4.0+ (`ipex`)         | 3.10+                     | Intel GPU | Experimental |
215
| **Ascend NPU** | 2.1.0+ (`torch_npu`)         | 3.10+                     | Ascend NPU | Experimental |
216
217
218
219
220
221
222
223
224
225

For each supported backend, follow the respective instructions below:

### Pre-requisites[[multi-backend-pre-requisites]]

To use bitsandbytes non-CUDA backends, be sure to install:

```
pip install "transformers>=4.45.1"
```
226

227
228
<hfoptions id="backend">
<hfoption id="AMD ROCm">
jiqing-feng's avatar
jiqing-feng committed
229

230
231
232
233
234
235
> [!WARNING]
> Pre-compiled binaries are only built for ROCm versions `6.1.0`/`6.1.1`/`6.1.2`/`6.2.0` and `gfx90a`, `gfx942`, `gfx1100` GPU architectures. [Find the pip install instructions here](#multi-backend-pip).
>
> Other supported versions that don't come with pre-compiled binaries [can be compiled for with these instructions](#multi-backend-compile).
>
> **Windows is not supported for the ROCm backend**; also not WSL2 to our knowledge.
jiqing-feng's avatar
jiqing-feng committed
236

237
> [!TIP]
238
> If you would like to install ROCm and PyTorch on bare metal, skip the Docker steps and refer to ROCm's official guides at [ROCm installation overview](https://rocm.docs.amd.com/projects/install-on-linux/en/latest/tutorial/install-overview.html#rocm-install-overview) and [Installing PyTorch for ROCm](https://rocm.docs.amd.com/projects/install-on-linux/en/latest/how-to/3rd-party/pytorch-install.html#using-wheels-package) (Step 3 of wheels build for quick installation). Special note: please make sure to get the respective ROCm-specific PyTorch wheel for the installed ROCm version, e.g. `https://download.pytorch.org/whl/nightly/rocm6.2/`!
jiqing-feng's avatar
jiqing-feng committed
239
240

```bash
241
242
243
244
# Create a docker container with latest ROCm image, which includes ROCm libraries
docker pull rocm/dev-ubuntu-22.04:6.1.2-complete
docker run -it --device=/dev/kfd --device=/dev/dri --group-add video rocm/dev-ubuntu-22.04:6.1.2-complete
apt-get update && apt-get install -y git && cd home
jiqing-feng's avatar
jiqing-feng committed
245

246
247
# Install pytorch compatible with above ROCm version
pip install torch --index-url https://download.pytorch.org/whl/rocm6.1/
248
```
249

250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
</hfoption>
<hfoption id="Intel CPU + GPU">

Compatible hardware and functioning `import intel_extension_for_pytorch as ipex` capable environment with Python `3.10` as the minimum requirement.

Please refer to [the official Intel installations instructions](https://intel.github.io/intel-extension-for-pytorch/index.html#installation?platform=cpu&version=v2.4.0%2bcpu&os=linux%2fwsl2) for guidance on how to pip install the necessary `intel_extension_for_pytorch` dependency.

</hfoption>
<hfoption id="Apple Silicon (MPS)">

> [!TIP]
> Apple Silicon support is still a WIP. Please visit and write us in [this Github Discussion space on coordinating the kickoff of MPS backend development](https://github.com/bitsandbytes-foundation/bitsandbytes/discussions/1340) and coordinate a community-led effort to implement this backend.

</hfoption>
</hfoptions>

### Installation

You can install the pre-built wheels for each backend, or compile from source for custom configurations.

#### Pre-built Wheel Installation (recommended)[[multi-backend-pip]]

<hfoptions id="platform">
<hfoption id="Linux">

```
# Note, if you don't want to reinstall BNBs dependencies, append the `--no-deps` flag!
pip install --force-reinstall 'https://github.com/bitsandbytes-foundation/bitsandbytes/releases/download/continuous-release_multi-backend-refactor/bitsandbytes-0.44.1.dev0-py3-none-manylinux_2_24_x86_64.whl'
```

</hfoption>
<hfoption id="Windows">

```
# Note, if you don't want to reinstall BNBs dependencies, append the `--no-deps` flag!
pip install --force-reinstall 'https://github.com/bitsandbytes-foundation/bitsandbytes/releases/download/continuous-release_multi-backend-refactor/bitsandbytes-0.44.1.dev0-py3-none-win_amd64.whl'
```

288
289
290
291
292
293
294
</hfoption>
<hfoption id="Ascend NPU">

Compatible hardware and functioning `import torch_npu` capable environment with Python `3.10` as the minimum requirement.

Please refer to [the official Ascend installations instructions](https://www.hiascend.com/document/detail/zh/Pytorch/60RC3/configandinstg/instg/insg_0001.html) for guidance on how to pip install the necessary `torch_npu` dependency.

295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
</hfoption>
<hfoption id="Mac">

> [!WARNING]
> bitsandbytes does not yet support Apple Silicon / Metal with a dedicated backend. However, the build infrastructure is in place and the below pip install will eventually provide Apple Silicon support as it becomes available on the `multi-backend-refactor` branch based on community contributions.

```
# Note, if you don't want to reinstall BNBs dependencies, append the `--no-deps` flag!
pip install --force-reinstall 'https://github.com/bitsandbytes-foundation/bitsandbytes/releases/download/continuous-release_multi-backend-refactor/bitsandbytes-0.44.1.dev0-py3-none-macosx_13_1_arm64.whl'
```

</hfoption>
</hfoptions>

#### Compile from Source[[multi-backend-compile]]

<hfoptions id="backend">
<hfoption id="AMD ROCm">

#### AMD GPU

bitsandbytes is fully supported from ROCm 6.1 onwards (currently in alpha release).

```bash
319
# Install bitsandbytes from source
320
# Clone bitsandbytes repo, ROCm backend is currently enabled on multi-backend-refactor branch
321
git clone -b multi-backend-refactor https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/
jiqing-feng's avatar
jiqing-feng committed
322

323
# Install dependencies
jiqing-feng's avatar
jiqing-feng committed
324
pip install -r requirements-dev.txt
325
326
327
328

# Compile & install
apt-get install -y build-essential cmake  # install build tools dependencies, unless present
cmake -DCOMPUTE_BACKEND=hip -S .  # Use -DBNB_ROCM_ARCH="gfx90a;gfx942" to target specific gpu arch
jiqing-feng's avatar
jiqing-feng committed
329
make
330
pip install -e .   # `-e` for "editable" install, when developing BNB (otherwise leave that out)
jiqing-feng's avatar
jiqing-feng committed
331
332
333
```

</hfoption>
334
<hfoption id="Intel CPU + GPU">
jiqing-feng's avatar
jiqing-feng committed
335

336
#### Intel CPU
jiqing-feng's avatar
jiqing-feng committed
337

338
339
> [!TIP]
> Intel CPU backend only supports building from source; for now, please follow the instructions below.
jiqing-feng's avatar
jiqing-feng committed
340

341
342
Similar to the CUDA case, you can compile bitsandbytes from source for Linux and Windows systems.

343
The below commands are for Linux. For installing on Windows, please adapt the below commands according to the same pattern as described [the section above on compiling from source under the Windows tab](#cuda-compile).
344
345

```
346
git clone --depth 1 -b multi-backend-refactor https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/
347
pip install intel_extension_for_pytorch
jiqing-feng's avatar
jiqing-feng committed
348
349
pip install -r requirements-dev.txt
cmake -DCOMPUTE_BACKEND=cpu -S .
350
351
make
pip install -e .   # `-e` for "editable" install, when developing BNB (otherwise leave that out)
jiqing-feng's avatar
jiqing-feng committed
352
353
```

354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
</hfoption>
<hfoption id="Ascend NPU">

#### Ascend NPU

> [!TIP]
> Ascend NPU backend only supports building from source; for now, please follow the instructions below.


```
# Install bitsandbytes from source
# Clone bitsandbytes repo, Ascend NPU backend is currently enabled on multi-backend-refactor branch
git clone -b multi-backend-refactor https://github.com/bitsandbytes-foundation/bitsandbytes.git && cd bitsandbytes/

# Install dependencies
pip install -r requirements-dev.txt

# Compile & install
apt-get install -y build-essential cmake  # install build tools dependencies, unless present
cmake -DCOMPUTE_BACKEND=npu -S .
make
pip install -e .   # `-e` for "editable" install, when developing BNB (otherwise leave that out)
```


379
380
381
</hfoption>
<hfoption id="Apple Silicon (MPS)">

382
383
#### Apple Silicon

384
385
WIP

jiqing-feng's avatar
jiqing-feng committed
386
387
</hfoption>
</hfoptions>