-
-
Notifications
You must be signed in to change notification settings - Fork 719
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
dask-worker
process remains alive after Nanny
exception on plugins=
#6320
Comments
Xfailing the test is needed for now due to dask/distributed#6320 . This should not have any major impact unless the user specifies a module that cannot be imported via `--pre-import`, in which case the worker process will hang indefinitely. Authors: - Peter Andreas Entschev (https://github.com/pentschev) Approvers: - https://github.com/jakirkham URL: #908
Probably not related, but FYI: #6357 |
This ticket is caused by this line distributed/distributed/core.py Line 325 in 63cdddd
We are now trying to close the server instance if an error occured to ensure all resources are properly cleaned up. I suspect that the worker is stuck trying to communicate to the scheduler or something (i.e. report=True). Fixing this is a bit nasty since we're calling |
What happened:
An exception on
Nanny
splugins=
doesn't cause thedask-worker
process to terminate after #5910 , instead it hangs indefinitely.What you expected to happen:
dask-worker
process to terminateMinimal Complete Verifiable Example:
Unfortunately, this isn't something that is implemented in
dask-worker
today, but a patch to reproduce this can be found below:To then reproduce, one would run the following CLIs:
Anything else we need to know?:
This behavior is relied upon in Dask-CUDA. #5910 caused https://github.com/rapidsai/dask-cuda/blob/7de73c72ca52239c6af87e483a20af3c8896bf0d/dask_cuda/tests/test_dask_cuda_worker.py#L220-L228 to hang indefinitely.
Environment:
cc @fjetter who authored #5910 .
The text was updated successfully, but these errors were encountered: