Dockerfile 3.37 KB
Newer Older
1
FROM nvidia/cuda:11.8.0-cudnn8-devel-ubuntu20.04
2
3
4
5
LABEL maintainer="Hugging Face"

ARG DEBIAN_FRONTEND=noninteractive

6
7
8
# Use login shell to read variables from `~/.profile` (to pass dynamic created variables between RUN commands)
SHELL ["sh", "-lc"]

9
10
11
# The following `ARG` are mainly used to specify the versions explicitly & directly in this docker file, and not meant
# to be used as arguments for docker build (so far).

12
ARG PYTORCH='2.1.0'
13
# (not always a valid torch version)
Yih-Dar's avatar
Yih-Dar committed
14
ARG INTEL_TORCH_EXT='2.1.0'
15
# Example: `cu102`, `cu113`, etc.
16
ARG CUDA='cu118'
17

18
RUN apt update
19
20
RUN apt install -y git libsndfile1-dev tesseract-ocr espeak-ng python3 python3-pip ffmpeg git-lfs
RUN git lfs install
21
22
RUN python3 -m pip install --no-cache-dir --upgrade pip

23
ARG REF=main
24
25
RUN git clone https://github.com/huggingface/transformers && cd transformers && git checkout $REF

26
27
28
29
30
31
32
# TODO: Handle these in a python utility script
RUN [ ${#PYTORCH} -gt 0 -a "$PYTORCH" != "pre" ] && VERSION='torch=='$PYTORCH'.*' ||  VERSION='torch'; echo "export VERSION='$VERSION'" >> ~/.profile
RUN echo torch=$VERSION
# `torchvision` and `torchaudio` should be installed along with `torch`, especially for nightly build.
# Currently, let's just use their latest releases (when `torch` is installed with a release version)
# TODO: We might need to specify proper versions that work with a specific torch version (especially for past CI).
RUN [ "$PYTORCH" != "pre" ] && python3 -m pip install --no-cache-dir -U $VERSION torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/$CUDA || python3 -m pip install --no-cache-dir -U --pre torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/nightly/$CUDA
33

34
RUN python3 -m pip install --no-cache-dir -U tensorflow==2.13 protobuf==3.20.3 tensorflow_text tensorflow_probability
35
36
37

RUN python3 -m pip install --no-cache-dir -e ./transformers[dev,onnxruntime]

38
RUN python3 -m pip uninstall -y flax jax
39

Yih-Dar's avatar
Yih-Dar committed
40
RUN python3 -m pip install --no-cache-dir intel_extension_for_pytorch==$INTEL_TORCH_EXT -f https://developer.intel.com/ipex-whl-stable-cpu
41

42
RUN python3 -m pip install --no-cache-dir git+https://github.com/facebookresearch/detectron2.git pytesseract
43
44
RUN python3 -m pip install -U "itsdangerous<2.1.0"

45
46
RUN python3 -m pip install --no-cache-dir git+https://github.com/huggingface/accelerate@main#egg=accelerate

47
48
RUN python3 -m pip install --no-cache-dir git+https://github.com/huggingface/peft@main#egg=peft

49
# Add bitsandbytes for mixed int8 testing
50
RUN python3 -m pip install --no-cache-dir bitsandbytes
51

Marc Sun's avatar
Marc Sun committed
52
# Add auto-gptq for gtpq quantization testing
53
RUN python3 -m pip install --no-cache-dir auto-gptq --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
Marc Sun's avatar
Marc Sun committed
54

55
56
57
# Add einops for additional model testing
RUN python3 -m pip install --no-cache-dir einops

58
59
60
# Add autoawq for quantization testing
RUN python3 -m pip install --no-cache-dir autoawq

Marc Sun's avatar
Marc Sun committed
61
62
# For bettertransformer + gptq 
RUN python3 -m pip install --no-cache-dir git+https://github.com/huggingface/optimum@main#egg=optimum
63

64
65
# For video model testing
RUN python3 -m pip install --no-cache-dir decord av==9.2.0
66

67
68
# For `dinat` model
RUN python3 -m pip install --no-cache-dir natten -f https://shi-labs.com/natten/wheels/$CUDA/
69

70
71
72
# When installing in editable mode, `transformers` is not recognized as a package.
# this line must be added in order for python to be aware of transformers.
RUN cd transformers && python3 setup.py develop