For vLLM, add support for `vllm serve`. Multi-node Inference can be done this way. vLLM docs - [Distributed Inference and Serving](https://docs.vllm.ai/en/stable/serving/distributed_serving.html?h=distribute)