Replies: 1 comment 4 replies
-
hi @rampa3 - I'm totally in favor of this and help on this would be much appreciated. If you open up the PR we can work out the details along the way. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I have been experimenting with producing CPU-only builds of the Python LocalAI backends using CPU only version of Torch (
+cpu
) from Torch's repository (--extra-index-url https://download.pytorch.org/whl/cpu
to prevent pulling CUDA dependences into CPU builds as Torch from PyPi does, since it is just a copy of the CUDA Torch), and in my opinion, at least some of them might be worth to have built officially. Don't want just outright try to work on a PR for something no one but me wants, so I am opening this discussion for it first.So far, I have these results at producing pure-CPU builds: (exllama2 is omitted, as it is a CUDA-only backend)
rfdetr-base
model from gallery is not valid (has official build, but it is being built against CUDA Torch from PyPi - the image is very big for no reason for CPU usage)For the non-working ones or untested ones, I plan to return to over the time when I have a while, attempting to get them all working.
If anyone else would want to also attempt working on these, my experimental (and probably a bit messy) CPU branch is available here (might not be always up to date): https://github.com/rampa3/LocalAI/tree/python_backends_cpu_build_tweaks
I look forward to any further discussions about the possibility of CPU builds of Python backends.
Beta Was this translation helpful? Give feedback.
All reactions