-
Hello, I was wondering if there is a simple way to use Thanks! |
Beta Was this translation helpful? Give feedback.
Answered by
slaren
Aug 30, 2024
Replies: 1 comment 1 reply
-
|
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
mjkpolo
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
test-backend-ops
does not support tensor parallelism. As far as I know yes, only llama.cpp supports it. However, all you need to do use it is to allocate the weights in aggml_backend_cuda_split_buffer_type
.