@@ -5,7 +5,7 @@ RUN mkdir -p /tgi
55# Fetch the optimum-neuron sources directly to avoid relying on pypi deployments
66FROM alpine AS optimum-neuron
77RUN mkdir -p /optimum-neuron
8- ADD https://github.com/huggingface/optimum-neuron/archive/refs/tags/v0.0.28 .tar.gz /optimum-neuron/sources.tar.gz
8+ ADD https://github.com/huggingface/optimum-neuron/archive/refs/tags/v0.1.0 .tar.gz /optimum-neuron/sources.tar.gz
99RUN tar -C /optimum-neuron -xf /optimum-neuron/sources.tar.gz --strip-components=1
1010
1111# Build cargo components (adapted from TGI original Dockerfile)
@@ -108,10 +108,10 @@ RUN wget -qO - https://apt.repos.neuron.amazonaws.com/GPG-PUB-KEY-AMAZON-AWS-NEU
108108# Install neuronx packages
109109RUN apt-get update -y \
110110 && apt-get install -y --no-install-recommends \
111- aws-neuronx-dkms=2.18.20 .0 \
112- aws-neuronx-collectives=2.22.33 .0-d2128d1aa \
113- aws-neuronx-runtime-lib=2.22.19 .0-5856c0b42 \
114- aws-neuronx-tools=2.19.0 .0 \
111+ aws-neuronx-dkms=2.19.64 .0 \
112+ aws-neuronx-collectives=2.23.135 .0-3e70920f2 \
113+ aws-neuronx-runtime-lib=2.23.112 .0-9b5179492 \
114+ aws-neuronx-tools=2.20.204 .0 \
115115 libxml2 \
116116 && rm -rf /var/lib/apt/lists/* \
117117 && apt-get clean
@@ -120,16 +120,16 @@ ENV PATH="/opt/bin/:/opt/aws/neuron/bin:${PATH}"
120120
121121# Install manually torch CPU version to avoid pulling CUDA
122122RUN pip3 install \
123- torch==2.1.2 \
124- torchvision==0.16.2 \
123+ torch==2.5.1 \
124+ torchvision==0.20.1 \
125125 --index-url https://download.pytorch.org/whl/cpu
126126
127127RUN pip3 install \
128- neuronx-cc==2.15.143 .0 \
129- torch-neuronx==2.1.2.2.3.2 \
130- transformers-neuronx==0.12.313 \
131- neuronx-distributed==0.9.0 \
132- libneuronxla==2.0.5347 .0 \
128+ neuronx-cc==2.16.372 .0 \
129+ torch-neuronx==2.5. 1.2.4.0 \
130+ transformers-neuronx==0.13.322 \
131+ neuronx-distributed==0.10.1 \
132+ libneuronxla==2.1.681 .0 \
133133 --extra-index-url=https://pip.repos.neuron.amazonaws.com
134134
135135# Install HuggingFace packages
0 commit comments