You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The bug is probably caused by an incorrect order among Pod Readiness, FUSE mount point readiness and CSI Plugin's recover() func logic.
Currently, CSI plugin recovers broken mount points only if it detects any FUSE container restarts and become ready. However, it is possible that the execution order goes like:
Alluxio FUSE Pod restarts
Alluxio FUSE Pod ready
CSI detects container restart and recovers broken mount points (At this time, Alluxio FUSE mount point is not ready, so nothing happened)
Alluxio FUSE mount point ready
CSI Plugin won't retry Step 1 to 4 because it detects no container restarts
The bug is probably caused by an incorrect order among Pod Readiness, FUSE mount point readiness and CSI Plugin's recover() func logic.
Currently, CSI plugin recovers broken mount points only if it detects any FUSE container restarts and become ready. However, it is possible that the execution order goes like:
Alluxio FUSE Pod restarts
Alluxio FUSE Pod ready
CSI detects container restart and recovers broken mount points (At this time, Alluxio FUSE mount point is not ready, so nothing happened)
Alluxio FUSE mount point ready
CSI Plugin won't retry Step 1 to 4 because it detects no container restarts
Can scanning /proc/self/mountinfo solve this issue?
What is your environment(Kubernetes version, Fluid version, etc.)
Describe the bug
FUSE Recovery failed when using AlluxioRuntime.
What you expect to happen:
Alluxio FUSE should be successfully recovered after deleting the FUSE pod.
How to reproduce it
Simply run the e2e script in #2477 can reproduce this bug.
Additional Information
The text was updated successfully, but these errors were encountered: