Memory Issues with Sparse Vectors in XGBoost4j-Spark: Disabling Sparse-to-Dense Conversion #11467
Unanswered
stepanov1997
asked this question in
Q&A
Replies: 2 comments 2 replies
-
cc @wbo4958 |
Beta Was this translation helpful? Give feedback.
0 replies
-
Hey @trivialfis and @wbo4958, the lack of proper handling for Can we expect this issue to be fixed in the near future, or should I consider giving up on distributed training for now? |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi everyone,
I’m using sparse vectors with about 10 features out of a possible 50 million. However, the conversion to dense vectors is causing heap exhaustion. Is there a way to disable the sparse-to-dense conversion?
Right now, I can’t even train on a small batch of vectors without running into memory issues, but I ultimately need to train on 200 million rows.
Any help would be greatly appreciated. I’m using XGBoost4j-Spark version 3.0.0 with the Java.
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions