-
Notifications
You must be signed in to change notification settings - Fork 311
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Forward merge branch-24.06 into branch-24.08 #4489
Forward merge branch-24.06 into branch-24.08 #4489
Conversation
…#4475) Addresses rapidsai#4474 Currently `openmpi=5.0.3-hfd7b305_105` is blocking our CI `cpp_build` job. Most likely introduced by this PR: conda-forge/openmpi-feedstock#158 This PR will unblock cugraph development until the issues are fixed. Once that happens, the version pinning should be removed.
…sai#4464) PyTorch 2.2+ is incompatible with the NCCL version on our containers. Normally, this would not be an issue, but there is a bug in CuPy that loads the system NCCL instead of the user NCCL. This PR binds the PyTorch test dependency version to get around this issue. --------- Co-authored-by: Bradley Dice <[email protected]> Co-authored-by: Ralph Liu <[email protected]> Co-authored-by: James Lamb <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
👍
@nv-rliu I think this should be renamed "Forward merge branch-24.06 into branch-24.08 |
ci/build_wheel.sh
Outdated
PARALLEL_LEVEL=$(python -c \ | ||
"from math import ceil; from multiprocessing import cpu_count; print(ceil(cpu_count()/2))") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
we recently added a PARALLEL_LEVEL
environment variable to the rapids-configure-sccache
script below:
That script is source
d earlier in this file.
Therefore you could simply use the value defined from that script instead of redefining it here.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great. Just wanted to be aggressive here to rule out the impact of hyperthreading. Using all 64 logical cores to compile cugraph did fail at times on my workstation (w/ Threadripper 3975WX).
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't think hyperthreading is a factor on these CI machines -- we get the "real" number of cores, afaik. We should be safe to remove this. Let's wait to push until after CI runs on the current commit (I'm testing something else at the moment).
I'm going to go ahead and trigger a merge. Builds all succeeded. |
/merge |
Replaces #4476