Replies: 4 comments 2 replies
-
Hey @al3rez! I'm here to help you with any bugs, questions, or contributions you have. Let's tackle this together while we wait for a human maintainer. To use the remote URL for the from langchain_community.llms.vllm import VLLM
# Example remote URL for the model
remote_model_url = "https://huggingface.co/remote-model-url"
# Create an instance of VLLM with the remote URL
llm = VLLM(
model=remote_model_url,
tensor_parallel_size=4, # Adjust based on your GPU setup
trust_remote_code=True, # Set to True if you trust the remote code
max_new_tokens=128,
top_k=10,
top_p=0.95,
temperature=0.8,
) Regarding whether to use a reverse proxy or a better approach, using a reverse proxy can be beneficial for load balancing, SSL termination, and other features. If you are deploying the model in a production environment, a reverse proxy like Nginx or Traefik could be a good choice. Alternatively, you can use the from langchain_community.llms import OpenLLM
# Example remote server URL
server_url = 'http://localhost:3000'
# Create an instance of OpenLLM with the remote server URL
llm = OpenLLM(server_url=server_url)
llm.invoke("What is the difference between a duck and a goose?") This approach allows you to leverage the |
Beta Was this translation helpful? Give feedback.
-
how can i do bind_tools with this? |
Beta Was this translation helpful? Give feedback.
-
@al3rez Hey, did you by any chance get a solution for this? |
Beta Was this translation helpful? Give feedback.
-
This is useful: https://python.langchain.com/docs/integrations/providers/runpod/ |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Checked other resources
Commit to Help
Example Code
Description
I'm running vllm on runpod 4x GPU using HuggingFace and vllm (casperhansen/llama-3-70b-instruct-awq) how can I use the remote URL for the VLLM() to create a llm? I've done it using chainlit + requests, but I want to be able to do it using
Any ideas?
System Info
System Information
Package Information
Packages not installed (Not Necessarily a Problem)
The following packages were not found:
Beta Was this translation helpful? Give feedback.
All reactions