-
Notifications
You must be signed in to change notification settings - Fork 69
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error 'Volume xxx Busy, status: UnmountUnderProgress' in cstor-csi-node pod #405
Comments
Can you try out with the latest release of OpenEBS 3.1? Suspecting it might have got fixed via cstor-csi#169 |
Hello @mittachaitu , sadly it is still happening with OpenEBS v3.1.0 (and Kubernetes v1.21.9). |
Hi @Ab-hishek as I stated in the issue description, we have only seen this problem when reusing a existing PVs. If I remember correctly (it's been some time), the error appeared during long running tests where StatefulSets that use cStor volumes were created and destroyed all over again many times. |
I'm seeing this with cstor-operator 3.3.0, also with StatefulSets rescheduling pods. I can work around it by killing the CSI node, but it seems like a legitimate race condition around
|
Deleting pod |
Kubernetes version: 1.21.6
OpenEBS versions: 3.0.0 (also seen in 2.11.x)
We are seeing this problem happen every now and then when reusing previously existing PVs.
The Pod that tries to reuse the PV will get stuck in ContainerCreating status.
Kubelet logs will display an error when requesting CSI to mount the volume:
The openebs-cstor-csi-node-xxxx Pod in that node is logging this error:
It seems it can't mount the volume and reports the same error over and over in every Kubelet request.
We have seen that if we delete the openebs-cstor-csi-node-xxxx Pod, the new openebs-cstor-csi-node-yyyy Pod will be able to mount the volume the first time Kubelet requests it. It looks like The old Pod was stuck for some reason, but the new one is able to carry out the mounting without any problems.
These are the logs of the new openebs-cstor-csi-node-yyyy:
Looking at openebs-cstor-csi-node-xxxx Pod logs again (the one that got stuck) we can see that the last time it was releasing that volume, an error was logged:
Wider log containing that error:
The text was updated successfully, but these errors were encountered: