-
Notifications
You must be signed in to change notification settings - Fork 921
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEA] Pass cudaStreamPerThread
to numba/CuPy kernels
#5922
Comments
Some relevant issues listed below: xref: rapidsai/dask-cuda#96 Also some relevant PRs below: xref: rapidsai/rmm#480 |
This issue has been marked rotten due to no recent activity in the past 90d. Please close this issue if no further response or action is needed. Otherwise, please respond with a comment indicating any updates or changes to the original issue and/or confirm this issue still needs to be addressed. |
cc @pentschev (for awareness) |
@shwina Is this intended to be unconditional, or only when requested by the user? We will also need to start passing streams down to libcudf everywhere in order to support this. |
My take on this is that we could begin with an opt-in solution, like we've done for CuPy with |
Yeah agreed. We probably want to support both the current mode and PTDS for some time. There are likely different issues we will run into. So having an escape hatch to old behavior is quite useful. |
Memory allocations should already use PTDS since both numba and CuPy allocate memory using RMM. Kernels on the other hand, may explicitly need to be passed the
cudaStreamPerThread
stream handle.cc: @jakirkham @kkraus14
The text was updated successfully, but these errors were encountered: