diff --git a/docker/presets/inference/llama-2/Dockerfile b/docker/presets/inference/llama-2/Dockerfile index 641d158bc..285cb122a 100644 --- a/docker/presets/inference/llama-2/Dockerfile +++ b/docker/presets/inference/llama-2/Dockerfile @@ -20,7 +20,7 @@ WORKDIR /workspace/llama RUN ["/bin/bash", "-c", "sed -i $'/torch.distributed.init_process_group(\"nccl\")/c\\ import datetime\\\n torch.distributed.init_process_group(\"nccl\", timeout=datetime.timedelta(days=365*100))' /workspace/llama/llama/generation.py"] RUN pip install -e . -RUN pip install torch==2.1.0 fastapi==0.103.2 pydantic==1.10.9 gputil==1.4.0 +RUN pip install torch==2.1.0 fastapi==0.109.1 pydantic==1.10.9 gputil==1.4.0 RUN pip install 'uvicorn[standard]' ARG WEIGHTS_PATH diff --git a/presets/tuning/tfs/requirements.txt b/presets/tuning/tfs/requirements.txt index 091e6d21f..9848f3e67 100644 --- a/presets/tuning/tfs/requirements.txt +++ b/presets/tuning/tfs/requirements.txt @@ -3,7 +3,7 @@ peft==0.8.2 transformers==4.38.2 torch==2.2.0 accelerate==0.27.2 -fastapi==0.103.2 +fastapi==0.109.1 pydantic==1.10.9 uvicorn[standard]==0.23.2 bitsandbytes==0.42.0