-
-
Notifications
You must be signed in to change notification settings - Fork 8.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conflict with a CUDA-11 PyTorch installation #10729
Comments
Thank you for raising an issue. I don't know what is the right solution either. As for a workaround, one can either use conda/mamba, or install xgboost without dependency ( The binary wheel itself can use |
May I ask what specific problem you ran into? |
After doing a PyTorch distributed mode init, when doing a
It got fixed after removing |
XGBoost for Python depends on
nvidia-nccl-cu12
, which is for CUDA 12. I have a PyTorch 2.4.0 installation for CUDA 11.8, but when I use distributed mode, PyTorch picks up on the one installed by XGBoost for NCCL and it gives me problems for my environment.My workaround is to install the CPU-only version of XGBoost. However, I still want to use XGBoost with CUDA support. It'd be nice if I could use it with
nvidia-nccl-cu11
instead. Not sure what the solution could be (maybe optional groups of dependencies for XGBoost, such as acu11
one, etc; or a different package). Note this could be a future problem when CUDA 13 comes out as well.The text was updated successfully, but these errors were encountered: