Replies: 1 comment 2 replies
-
Models don't get converted to C++. They just get converted to the format the project can use (GGUF). The easiest way to get started is to just download a model that's already been converted to GGUF format: https://huggingface.co/TheBloke 7B models are pretty small and should run on low memory systems. 33+B models you probably have to have at least 32GB RAM. For 65B and 70B you need 64GB RAM. You can get an idea of how much memory will be required from the file sizes: generally you'll need memory available at least equal to the size of the model file plus a few gigs extra for stuff like context. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Love this project!
I am currently only writing Python and don't really understand what you did here, can you recommend some resources for converting models to C++ and to use hardware acceleration on Mac?
Thank you in advance!
Beta Was this translation helpful? Give feedback.
All reactions