LoRa @ Inference Time. #1833
Unanswered
Clarie-Starling
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
Despite it not being said do the Model Configs in .env.local Support the OpenAI lora: option?
during a chat/completion invoke? Or even static in the server config?
I.e
MODELS=`[
{
"name": "Local microsoft/Phi-3-mini-4k-instruct-gguf",
"tokenizer": "microsoft/Phi-3-mini-4k-instruct-gguf",
"lora": "My/LoRa/Path/backwards.safetensors",
This is a big need due to apps needing AI personality is becoming more and more prevalent (System prompts don't cut it anymore)
Beta Was this translation helpful? Give feedback.
All reactions