Makefile 1.79 KB
Newer Older
1
transformers_commit := 2b57aa18da658e7d2f42ef6bd5b56751af582fef
2
flash_att_commit := 4d87e4d875077ad9efd25030efa4ab0ba92c19e1
3

Olivier Dehaene's avatar
Init  
Olivier Dehaene committed
4
gen-server:
Nicolas Patry's avatar
Nicolas Patry committed
5
	# Compile protos
6
	pip install grpcio-tools==1.51.1 mypy-protobuf==3.4.0 'types-protobuf>=3.20.4' --no-cache-dir
7
	mkdir text_generation_server/pb || true
8
9
	python -m grpc_tools.protoc -I../proto --python_out=text_generation_server/pb \
		--grpc_python_out=text_generation_server/pb --mypy_out=text_generation_server/pb ../proto/generate.proto
10
11
	find text_generation_server/pb/ -type f -name "*.py" -print0 -exec sed -i -e 's/^\(import.*pb2\)/from . \1/g' {} \;
	touch text_generation_server/pb/__init__.py
Olivier Dehaene's avatar
Init  
Olivier Dehaene committed
12

Olivier Dehaene's avatar
v0.1.0  
Olivier Dehaene committed
13
install-transformers:
14
	# Install specific version of transformers with custom cuda kernels
15
16
	pip uninstall transformers -y || true
	rm -rf transformers || true
17
18
	git clone https://github.com/OlivierDehaene/transformers.git
	cd transformers && git checkout $(transformers_commit)
Olivier Dehaene's avatar
v0.1.0  
Olivier Dehaene committed
19
	cd transformers && python setup.py install
Olivier Dehaene's avatar
Init  
Olivier Dehaene committed
20

21
22
23
24
25
26
27
28
29
install-flash-attention:
	# Install specific version of flash attention
	pip install packaging
	pip uninstall flash_attn rotary_emb dropout_layer_norm -y || true
	rm -rf flash-attention || true
	git clone https://github.com/HazyResearch/flash-attention.git
	cd flash-attention && git checkout $(flash_att_commit)
	cd flash-attention && python setup.py install && cd csrc/layer_norm && python setup.py install && cd ../rotary && python setup.py install

Olivier Dehaene's avatar
v0.1.0  
Olivier Dehaene committed
30
31
install-torch:
	# Install specific version of torch
32
	pip install torch --extra-index-url https://download.pytorch.org/whl/cu118 --no-cache-dir
Olivier Dehaene's avatar
Init  
Olivier Dehaene committed
33

OlivierDehaene's avatar
OlivierDehaene committed
34
install: gen-server install-torch install-transformers
Nicolas Patry's avatar
Nicolas Patry committed
35
36
	pip install pip --upgrade
	pip install -e . --no-cache-dir
Olivier Dehaene's avatar
Init  
Olivier Dehaene committed
37

Nicolas Patry's avatar
Nicolas Patry committed
38
run-dev:
39
	SAFETENSORS_FAST_GPU=1 python -m torch.distributed.run --nproc_per_node=2 text_generation_server/cli.py serve bigscience/bloom-560m --sharded