Could not install TabbyML hosted in K8s & integrated with Cloud Based APIs like OpenAI #3109
Replies: 3 comments 1 reply
-
Hello, it appears that your configuration file is not in the |
Beta Was this translation helpful? Give feedback.
-
Hi @wsxiaoys, Thanks for reverting back. Just to give a little background. I'm trying to self-host aforesaid open-source docker image version of tabbyml onto my K8s cluster. As a part of self-hosting I have used following K8s resources:
K8s installation of any product never have file with name config.toml. Pls guide how to make tabbyml work while self-hosting. Just FYI, my pods are up and running but still instances are not working as it expects the llama-server model to be present in my k8s infra instead of picking up from openAI model. Many thanks |
Beta Was this translation helpful? Give feedback.
-
@siddharthgaur2590 how do you mount the configMap into your pod? looks like its not formatted in maybe your should create a configMap by a file containing toml configuration described in https://tabby.tabbyml.com/docs/administration/model/, and then mount the configMap into the tabby pod under |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Describe the bug
Trying to self host the TabbyML on K8s using Cloud Based API like openAI ones which have the GPUs capability but facing below error.
WARN llama_cpp_server::supervisor: crates/llama-cpp-server/src/supervisor.rs:98: llama-server exited with status code 127, args:
Command { std: "/opt/tabby/bin/llama-server" "-m" "/data/models/TabbyML/Nomic-Embed-Text/ggml/model.gguf" "--cont-batching" "--port" "30888" "-np" "1" "--log-disable" "--ctx-size" "4096" "-ngl" "9999" "--embedding" "--ubatch-size" "4096", kill_on_drop: true }
2024-09-05T08:16:26.589Z | 2024-09-05T08:16:26.589532Z WARN llama_cpp_server::supervisor: crates/llama-cpp-server/src/supervisor.rs:110: : /opt/tabby/bin/llama-server: error while loading shared libraries: libcuda.so.1: cannot open shared object file: No such file or directory
Information about your version
tabbyml/tabby:20240826
Information about your GPU
Using the config mentioned herein: https://tabby.tabbyml.com/docs/references/models-http-api/openai/
Additional context
Following config are used to direct the tabbyML to use GPU based models present in OpenAI platform.
Beta Was this translation helpful? Give feedback.
All reactions