Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LocalCUDACluster's memory limit: None means no limit #943

Merged

Conversation

madsbk
Copy link
Member

@madsbk madsbk commented Jul 4, 2022

Use a regular dict when creating a LocalCUDACluster with no host and no device memory limit.

Currently, setting device_memory_limit=None translate into the total available GPU memory. However, in some cases DeviceHostFile overestimate the GPU memory usage, which can trigger spilling even though device_memory_limit=None.

@madsbk madsbk added bug Something isn't working 2 - In Progress Currently a work in progress non-breaking Non-breaking change labels Jul 4, 2022
@github-actions github-actions bot added the python python code needed label Jul 4, 2022
@codecov-commenter
Copy link

codecov-commenter commented Jul 4, 2022

Codecov Report

❗ No coverage uploaded for pull request base (branch-22.08@b671e8d). Click here to learn what that means.
The diff coverage is n/a.

❗ Current head 45e246b differs from pull request most recent head c1a012d. Consider uploading reports for the commit c1a012d to get more accurate results

@@              Coverage Diff               @@
##             branch-22.08    #943   +/-   ##
==============================================
  Coverage                ?   0.00%           
==============================================
  Files                   ?      16           
  Lines                   ?    2105           
  Branches                ?       0           
==============================================
  Hits                    ?       0           
  Misses                  ?    2105           
  Partials                ?       0           

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update b671e8d...c1a012d. Read the comment docs.

@madsbk madsbk changed the title LocalCUDACluster: no memory limits means *no* limit LocalCUDACluster's memory limit: no means no Jul 4, 2022
@madsbk madsbk marked this pull request as ready for review July 4, 2022 14:20
@madsbk madsbk requested a review from a team as a code owner July 4, 2022 14:20
@madsbk madsbk added 3 - Ready for Review Ready for review by team and removed 2 - In Progress Currently a work in progress labels Jul 4, 2022
Copy link
Member

@pentschev pentschev left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM, thanks @madsbk .

@pentschev pentschev changed the title LocalCUDACluster's memory limit: no means no LocalCUDACluster's memory limit: None means no limit Jul 4, 2022
@pentschev
Copy link
Member

@gpucibot merge

@rapids-bot rapids-bot bot merged commit 8dba7d1 into rapidsai:branch-22.08 Jul 4, 2022
@madsbk madsbk deleted the CUDAWorker_no_device_limit branch January 24, 2023 12:23
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
3 - Ready for Review Ready for review by team bug Something isn't working non-breaking Non-breaking change python python code needed
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants