Sharing my vLLM Docker Container Image #1454
samos123
announced in
Show and tell
Replies: 1 comment 1 reply
-
Can you explain how I can interact with the docker container with Langchain for instance? Or can I use the HF text-inference framework and call this as an API endpoint? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Repo: https://github.com/substratusai/vllm-docker
This container image runs the OpenAI API server of vLLM.
Image URL:
ghcr.io/substratusai/vllm
Quickstart
Deploy Mistral 7B Instruct:
Configuration Options
The following configuration options are available by using environment
variables:
hf_org/model
or utilize a path to point to a local model. Example value: mistralai/Mistral-7B-Instruct-v0.1The container image automatically detects the number of GPUs and sets
--tensor-parallel-size
to be equal to number of GPUs available. Thegpu-count.py
script is used to detect number of GPUs.Building
Beta Was this translation helpful? Give feedback.
All reactions