Skip to content

Commit 5561681

Browse files
authored
[CI] add kvcache-connector dependency definition and add into CI build (#18193)
Signed-off-by: Peter Pan <Peter.Pan@daocloud.io>
1 parent fbd62d8 commit 5561681

File tree

3 files changed

+14
-1
lines changed

3 files changed

+14
-1
lines changed

.buildkite/release-pipeline.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ steps:
5252
queue: cpu_queue_postmerge
5353
commands:
5454
- "aws ecr-public get-login-password --region us-east-1 | docker login --username AWS --password-stdin public.ecr.aws/q9t5s3a7"
55-
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.8.1 --tag public.ecr.aws/q9t5s3a7/vllm-release-repo:$BUILDKITE_COMMIT --target vllm-openai --progress plain -f docker/Dockerfile ."
55+
- "DOCKER_BUILDKIT=1 docker build --build-arg max_jobs=16 --build-arg USE_SCCACHE=1 --build-arg GIT_REPO_CHECK=1 --build-arg CUDA_VERSION=12.8.1 --build-arg INSTALL_KV_CONNECTORS=true --tag public.ecr.aws/q9t5s3a7/vllm-release-repo:$BUILDKITE_COMMIT --target vllm-openai --progress plain -f docker/Dockerfile ."
5656
- "docker push public.ecr.aws/q9t5s3a7/vllm-release-repo:$BUILDKITE_COMMIT"
5757

5858
- label: "Annotate release workflow"

docker/Dockerfile

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,4 @@
1+
12
# The vLLM Dockerfile is used to construct vLLM image that can be directly used
23
# to run the OpenAI compatible server.
34

@@ -62,12 +63,16 @@ ARG PYTORCH_CUDA_NIGHTLY_INDEX_BASE_URL=https://download.pytorch.org/whl/nightly
6263
ARG PIP_KEYRING_PROVIDER=disabled
6364
ARG UV_KEYRING_PROVIDER=${PIP_KEYRING_PROVIDER}
6465

66+
# Flag enables build-in KV-connector dependency libs into docker images
67+
ARG INSTALL_KV_CONNECTORS=false
68+
6569
#################### BASE BUILD IMAGE ####################
6670
# prepare basic build environment
6771
FROM ${BUILD_BASE_IMAGE} AS base
6872
ARG CUDA_VERSION
6973
ARG PYTHON_VERSION
7074
ARG TARGETPLATFORM
75+
ARG INSTALL_KV_CONNECTORS=false
7176
ENV DEBIAN_FRONTEND=noninteractive
7277

7378
ARG DEADSNAKES_MIRROR_URL
@@ -276,6 +281,7 @@ RUN --mount=type=cache,target=/root/.cache/uv \
276281
FROM ${FINAL_BASE_IMAGE} AS vllm-base
277282
ARG CUDA_VERSION
278283
ARG PYTHON_VERSION
284+
ARG INSTALL_KV_CONNECTORS=false
279285
WORKDIR /vllm-workspace
280286
ENV DEBIAN_FRONTEND=noninteractive
281287
ARG TARGETPLATFORM
@@ -485,6 +491,7 @@ RUN mv mkdocs.yaml test_docs/
485491
# base openai image with additional requirements, for any subsequent openai-style images
486492
FROM vllm-base AS vllm-openai-base
487493
ARG TARGETPLATFORM
494+
ARG INSTALL_KV_CONNECTORS=false
488495

489496
ARG PIP_INDEX_URL UV_INDEX_URL
490497
ARG PIP_EXTRA_INDEX_URL UV_EXTRA_INDEX_URL
@@ -493,8 +500,13 @@ ARG PIP_EXTRA_INDEX_URL UV_EXTRA_INDEX_URL
493500
# Reference: https://github.com/astral-sh/uv/pull/1694
494501
ENV UV_HTTP_TIMEOUT=500
495502

503+
COPY requirements/kv_connectors.txt requirements/kv_connectors.txt
504+
496505
# install additional dependencies for openai api server
497506
RUN --mount=type=cache,target=/root/.cache/uv \
507+
if [ "$INSTALL_KV_CONNECTORS" = "true" ]; then \
508+
uv pip install --system -r requirements/kv_connectors.txt; \
509+
fi; \
498510
if [ "$TARGETPLATFORM" = "linux/arm64" ]; then \
499511
uv pip install --system accelerate hf_transfer 'modelscope!=1.15.0' 'bitsandbytes>=0.42.0' 'timm==0.9.10' boto3 runai-model-streamer runai-model-streamer[s3]; \
500512
else \

requirements/kv_connectors.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
lmcache

0 commit comments

Comments
 (0)