installation.md 2.19 KB
Newer Older
1
2
3
4
5
6
# Installation

This section explains how to install the CLI tool as well as installing TGI from source. **The strongly recommended approach is to use Docker, as it does not require much setup. Check [the Quick Tour](./quicktour) to learn how to run TGI with Docker.**

## Install CLI

Merve Noyan's avatar
Merve Noyan committed
7
You can use TGI command-line interface (CLI) to download weights, serve and quantize models, or get information on serving parameters. 
8

Merve Noyan's avatar
Merve Noyan committed
9
10
11
12
13
14
15
16
17
18
19
20
To install the CLI, you need to first clone the TGI repository and then run `make`.

```bash
git clone https://github.com/huggingface/text-generation-inference.git && cd text-generation-inference
make install
```

If you would like to serve models with custom kernels, run

```bash
BUILD_EXTENSIONS=True make install
```
21
22
23
24
25
26
27
28
29
30

## Local Installation from Source

Before you start, you will need to setup your environment, and install Text Generation Inference. Text Generation Inference is tested on **Python 3.9+**.

Text Generation Inference is available on pypi, conda and GitHub. 

To install and launch locally, first [install Rust](https://rustup.rs/) and create a Python virtual environment with at least
Python 3.9, e.g. using conda:

31
```bash
32
33
34
35
36
37
38
39
40
41
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh

conda create -n text-generation-inference python=3.9
conda activate text-generation-inference
```

You may also need to install Protoc.

On Linux:

42
```bash
43
44
45
46
47
48
49
50
51
PROTOC_ZIP=protoc-21.12-linux-x86_64.zip
curl -OL https://github.com/protocolbuffers/protobuf/releases/download/v21.12/$PROTOC_ZIP
sudo unzip -o $PROTOC_ZIP -d /usr/local bin/protoc
sudo unzip -o $PROTOC_ZIP -d /usr/local 'include/*'
rm -f $PROTOC_ZIP
```

On MacOS, using Homebrew:

52
```bash
53
54
55
56
57
brew install protobuf
```

Then run to install Text Generation Inference:

58
```bash
Merve Noyan's avatar
Merve Noyan committed
59
60
git clone https://github.com/huggingface/text-generation-inference.git && cd text-generation-inference
BUILD_EXTENSIONS=True make install
61
62
63
64
65
66
```

<Tip warning={true}>

On some machines, you may also need the OpenSSL libraries and gcc. On Linux machines, run:

67
```bash
68
69
70
71
72
73
74
sudo apt-get install libssl-dev gcc -y
```

</Tip>

Once installation is done, simply run:

75
```bash
76
77
78
79
make run-falcon-7b-instruct
```

This will serve Falcon 7B Instruct model from the port 8080, which we can query.