-
Notifications
You must be signed in to change notification settings - Fork 44
gpuOffload config parameter does not work #88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I also tried another syntax, similar to what's documented in lmsutudio-js:
This produces another error:
|
Please fix gpuOffload parameter or provide documentation how to really use it. |
This is a genuine bug in the SDK. The translation from the named "favorMainGpu" split strategy to the server's "priorityOrder" split configuration is not currently working correctly (without a clear client side workaround). |
Thank you for the response! You've mentioned
Can you provide more details about workaround? |
That was poor wording on my part - while the dynamic nature of Python means monkeypatching the SDK to fix the config translation is theoretically possible, actually doing so is sufficiently complicated that it doesn't really count as a viable workaround (and once I've worked out what is wrong in sufficient detail to describe what such a patch would look like, I'll have spun a new release with the translation fixed). |
Thank you for the information! BTW, meanwhile I've found a workaround - I'm just invoking LMS CLI executable for the loading, and it works as expected. |
Uh oh!
There was an error while loading. Please reload this page.
LM Studio version: 0.3.16 (Build 3)
Python package version: lmstudio==1.3.0
I've tried to write a custom model loading script using lmstudio-python API.
And one important thing is to specify custom gpuOffload.
I've tried to do it as it's documented here: https://lmstudio.ai/docs/python/llm-prediction/parameters
But it just doesn't have any effect.
Moreover, pylint checker gives an error:
I've tried to do some research, and found out that more correct way according to Python type hints is following:
But this this produces bunch of runtime errors.
The text was updated successfully, but these errors were encountered: