Skip to content

Commit 6255f29

Browse files
committed
vllm-sim references to llm-d-inference-sim
1 parent 35d9f4a commit 6255f29

File tree

20 files changed

+49
-49
lines changed

20 files changed

+49
-49
lines changed

.tekton/pipelinerun.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
apiVersion: tekton.dev/v1
22
kind: PipelineRun
33
metadata:
4-
name: vllm-sim
4+
name: llm-d-inference-sim
55
annotations:
66
pipelinesascode.tekton.dev/on-event: "[pull_request, push]"
77
pipelinesascode.tekton.dev/on-target-branch: "[main, dev]"

.version.json

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
{
22
"dev-version": "0.0.6",
3-
"dev-registry": "ghcr.io/llm-d/vllm-sim-dev",
3+
"dev-registry": "ghcr.io/llm-d/llm-d-inference-sim-dev",
44
"prod-version": "0.0.5",
5-
"prod-registry": "ghcr.io/llm-d/vllm-sim"
5+
"prod-registry": "ghcr.io/llm-d/llm-d-inference-sim"
66
}

Dockerfile

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -12,23 +12,23 @@ COPY go.sum go.sum
1212
RUN go mod download
1313

1414
# Copy the go source
15-
COPY cmd/vllm-sim/main.go cmd/cmd.go
15+
COPY cmd/llm-d-inference-sim/main.go cmd/cmd.go
1616
COPY . .
1717

1818
# Build
1919
# the GOARCH has not a default value to allow the binary be built according to the host where the command
2020
# was called. For example, if we call make image-build in a local env which has the Apple Silicon M1 SO
2121
# the docker BUILDPLATFORM arg will be linux/arm64 when for Apple x86 it will be linux/amd64. Therefore,
2222
# by leaving it empty we can ensure that the container and binary shipped on it will have the same platform.
23-
RUN CGO_ENABLED=0 GOOS=${TARGETOS:-linux} GOARCH=${TARGETARCH} go build -a -o bin/vllm-sim cmd/cmd.go
23+
RUN CGO_ENABLED=0 GOOS=${TARGETOS:-linux} GOARCH=${TARGETARCH} go build -a -o bin/llm-d-inference-sim cmd/cmd.go
2424

2525
# Use distroless as minimal base image to package the manager binary
2626
# Refer to https://github.com/GoogleContainerTools/distroless for more details
2727
FROM registry.access.redhat.com/ubi9/ubi:latest
2828
WORKDIR /
29-
COPY --from=builder /workspace/bin/vllm-sim /app/vllm-sim
29+
COPY --from=builder /workspace/bin/llm-d-inference-sim /app/llm-d-inference-sim
3030
USER 65532:65532
3131

32-
ENTRYPOINT ["/app/vllm-sim"]
32+
ENTRYPOINT ["/app/llm-d-inference-sim"]
3333

3434

Makefile

Lines changed: 12 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# Copyright 2025 The vLLM-Sim Authors.
1+
# Copyright 2025 The llm-d-inference-sim Authors.
22
#
33
# Licensed under the Apache License, Version 2.0 (the "License");
44
# you may not use this file except in compliance with the License.
@@ -12,31 +12,31 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
# Makefile for the vllm-sim project
15+
# Makefile for the llm-d-inference-sim project
1616

1717
CONTAINER_RUNTIME ?= docker
1818

19-
PACKAGE_VLLM_SIM = github.com/neuralmagic/vllm-sim/cmd/vllm-sim
20-
VLLM_SIM_NAME = vllm-sim/vllm-sim
19+
PACKAGE_VLLM_SIM = github.com/llm-d/llm-d-inference-sim/cmd/llm-d-inference-sim
20+
VLLM_SIM_NAME = llm-d-inference-sim/llm-d-inference-sim
2121
VLLM_SIM_TAG ?= 0.0.2
2222

23-
.PHONY: build-vllm-sim
24-
build-vllm-sim:
23+
.PHONY: build-llm-d-inference-sim
24+
build-llm-d-inference-sim:
2525
go build -o bin/ ${PACKAGE_VLLM_SIM}
2626

27-
.PHONY: build-vllm-sim-linux
28-
build-vllm-sim-linux:
27+
.PHONY: build-llm-d-inference-sim-linux
28+
build-llm-d-inference-sim-linux:
2929
CGO_ENABLED=0 GOOS=linux GOARCH=amd64 go build -o bin/linux/ ${PACKAGE_VLLM_SIM}
3030

31-
.PHONY: build-vllm-sim-image
32-
build-vllm-sim-image: build-vllm-sim-linux
33-
$(CONTAINER_RUNTIME) build --file build/vllm-sim.Dockerfile --tag ${VLLM_SIM_NAME}:${VLLM_SIM_TAG} ./bin/linux
31+
.PHONY: build-llm-d-inference-sim-image
32+
build-llm-d-inference-sim-image: build-llm-d-inference-sim-linux
33+
$(CONTAINER_RUNTIME) build --file build/llm-d-inference-sim.Dockerfile --tag ${VLLM_SIM_NAME}:${VLLM_SIM_TAG} ./bin/linux
3434

3535

3636
SHELL := /usr/bin/env bash
3737

3838
# Defaults
39-
PROJECT_NAME ?= vllm-sim
39+
PROJECT_NAME ?= llm-d-inference-sim
4040
DEV_VERSION ?= 0.0.1
4141
PROD_VERSION ?= 0.0.0
4242
IMAGE_TAG_BASE ?= ghcr.io/llm-d/$(PROJECT_NAME)

README.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -89,13 +89,13 @@ For more details see the <a href="https://docs.vllm.ai/en/stable/getting_started
8989
### Building
9090
To build a Docker image of the vLLM Simulator, run:
9191
```bash
92-
make build-vllm-sim-image
92+
make build-llm-d-inference-sim-image
9393
```
9494

9595
### Running
9696
To run the vLLM Simulator image under Docker, run:
9797
```bash
98-
docker run --rm --publish 8000:8000 ai-aware-router/vllm-sim:0.0.1 /ai-aware-router/vllm-sim --port 8000 --model "Qwen/Qwen2.5-1.5B-Instruct" --lora "tweet-summary-0,tweet-summary-1"
98+
docker run --rm --publish 8000:8000 ai-aware-router/llm-d-inference-sim:0.0.1 /ai-aware-router/llm-d-inference-sim --port 8000 --model "Qwen/Qwen2.5-1.5B-Instruct" --lora "tweet-summary-0,tweet-summary-1"
9999
```
100100
**Note:** The above command exposes the simulator on port 8000, and serves the Qwen/Qwen2.5-1.5B-Instruct model.
101101

@@ -104,13 +104,13 @@ docker run --rm --publish 8000:8000 ai-aware-router/vllm-sim:0.0.1 /ai-aware-rou
104104
### Building
105105
To build the vLLM simulator, run:
106106
```bash
107-
make build-vllm-sim
107+
make build-llm-d-inference-sim
108108
```
109109

110110
### Running
111111
To run the router in a standalone test environment, run:
112112
```bash
113-
./bin/vllm-sim --model my_model --port 8000
113+
./bin/llm-d-inference-sim --model my_model --port 8000
114114
```
115115

116116

build/vllm-sim.Dockerfile

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# Copyright 2025 The vLLM-Sim Authors.
1+
# Copyright 2025 The llm-d-inference-sim Authors.
22
#
33
# Licensed under the Apache License, Version 2.0 (the "License");
44
# you may not use this file except in compliance with the License.
@@ -14,8 +14,8 @@
1414

1515
FROM alpine:3.21
1616

17-
ADD --chmod=777 ./vllm-sim /vllm-sim/
17+
ADD --chmod=777 ./llm-d-inference-sim /llm-d-inference-sim/
1818

19-
WORKDIR /vllm-sim/
19+
WORKDIR /llm-d-inference-sim/
2020

21-
ENTRYPOINT ["/vllm-sim/vllm-sim"]
21+
ENTRYPOINT ["/llm-d-inference-sim/llm-d-inference-sim"]

cmd/vllm-sim/main.go renamed to cmd/llm-d-inference-sim/main.go

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
/*
2-
Copyright 2025 The vLLM-Sim Authors.
2+
Copyright 2025 The llm-d-inference-sim Authors.
33
44
Licensed under the Apache License, Version 2.0 (the "License");
55
you may not use this file except in compliance with the License.
@@ -23,8 +23,8 @@ import (
2323

2424
"k8s.io/klog/v2"
2525

26-
"github.com/neuralmagic/vllm-sim/cmd/signals"
27-
vllmsim "github.com/neuralmagic/vllm-sim/pkg/vllm-sim"
26+
"github.com/llm-d/llm-d-inference-sim/cmd/signals"
27+
vllmsim "github.com/llm-d/llm-d-inference-sim/pkg/llm-d-inference-sim"
2828
)
2929

3030
func main() {

cmd/signals/signals.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
/*
2-
Copyright 2025 The vLLM-Sim Authors.
2+
Copyright 2025 The llm-d-inference-sim Authors.
33
44
Licensed under the Apache License, Version 2.0 (the "License");
55
you may not use this file except in compliance with the License.

go.mod

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
module github.com/neuralmagic/vllm-sim
1+
module github.com/llm-d/llm-d-inference-sim
22

33
go 1.23.4
44

pkg/vllm-api/vllm-models.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
/*
2-
Copyright 2025 The vLLM-Sim Authors.
2+
Copyright 2025 The llm-d-inference-sim Authors.
33
44
Licensed under the Apache License, Version 2.0 (the "License");
55
you may not use this file except in compliance with the License.

0 commit comments

Comments
 (0)