Replies: 1 comment
-
Try this tutorial and see if it works for your model: ggml-org/llama.cpp#7712 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I want to run a BERT-like model on my CPU using GGML to see if it's faster than the PyTorch version. I can't find any documentation on how to use GGML. What's the easiest way of doing something like this?
Beta Was this translation helpful? Give feedback.
All reactions