Capacity Planning while using GPU based Indices #37632
Unanswered
rohitreddy1698
asked this question in
Q&A and General discussion
Replies: 3 comments 6 replies
-
I am using the milvus-2.4.9-gpu version of the docker image. |
Beta Was this translation helpful? Give feedback.
5 replies
-
@rohitreddy1698 I am currently working on implementing this feature, and if everything goes smoothly, I should be able to finish it by next week. Thank you for your patience. |
Beta Was this translation helpful? Give feedback.
1 reply
-
Hi @presburger , @yhmo Has there been any development in this area ? Thanks and Regards, |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello,
I am trying to compare the performances of GPU vs CPU based indices in Milvus and have Milvus setup on GKE for the same.
I have deployed Milvus using the Milvus Operator.
I am trying to do the benchmarking using the Zilliz Vector DB Bench tool : https://github.com/zilliztech/VectorDBBench
I have
4 : e2-highmen-16 nodes ( for other components )
30 : n1-highmen-8 nodes ( 25 for queyrnode and 5 for indexnode ) with 2 T4 GPUs ( 16 GiB memory ) per node
The data set I am using is 100M Laion dataset with 768 vector dimension. I am trying to test pure Search performance without any fiters.
I have a total of 25 * 2 * 16 = 800 GiB of GPU memory , which should be sufficient for GPU_IVF_FLAT and IVF_FLAT , but I am getting the following failed to deserialise error :
Please help me understand the sizes taken up by the GPU indices and help me get over this issue.
Beta Was this translation helpful? Give feedback.
All reactions