You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
What happened:
In an AKS environment, mounting a large NFS file share (12Ti, 11.45M files) to a pod causes significant delays, with the pod staying in "ContainerCreating" status for about 3 days. This occurs when syncing data via a linux vm on the nfs share, after sync the nfs share does mount in the pv and pvc correctly, pod does not start up due to operation on current pv already exists
What you expected to happen:
Pod starts with the volume attached immediately
How to reproduce it:
have a storage count
expose a file share under that storage account approx 1Ti with data with NFS
mount the disk under a linux vm
provision data via the linux vm (have considerable amount of files in the storage account eg 2.4M)
on aks provision the pv and attach the pvc in a namespace eg:
Warning FailedMount 2m24s (x857 over 28h) kubelet MountVolume.MountDevice failed for volume "pv-shared-home" : rpc error: code = Aborted desc = An operation with the given Volume ID pv-shared-home already exists
Anything else we need to know?:
Issue is occurs also after we have synced data (rsync to linux vm in azure, with nfs share mounted) into this NFS share.
Once the PV has been mounted and no operations are ongoing the PV and PVC can be removed. 2nd run the pvc is immediately mounted and available. Even if we switch from cluster its fine.
Environment:
CSI Driver version: v1.29.5
Kubernetes version (use kubectl version): 1.28.3
OS (e.g. from /etc/os-release): AKSUbuntu-2204gen2containerd-202405.20.0
Kernel (e.g. uname -a):
Install tools:
Others:
The text was updated successfully, but these errors were encountered:
@ashwajce could you provide the kubelet logs on that node in problem? have you set any securityContext in pod? this issue could be related to slow chown operation if you set fsGroup in securityContext in pod, one workaround is set fsGroupChangePolicy: None in pv
fsGroupChangePolicy: indicates how volume's ownership will be changed by the driver, pod securityContext.fsGroupChangePolicy is ignored
k8s-ci-robot
added
lifecycle/rotten
Denotes an issue or PR that has aged beyond stale and will be auto-closed.
and removed
lifecycle/stale
Denotes an issue or PR has remained open with no activity and has become stale.
labels
Nov 4, 2024
What happened:
In an AKS environment, mounting a large NFS file share (12Ti, 11.45M files) to a pod causes significant delays, with the pod staying in "ContainerCreating" status for about 3 days. This occurs when syncing data via a linux vm on the nfs share, after sync the nfs share does mount in the pv and pvc correctly, pod does not start up due to operation on current pv already exists
What you expected to happen:
Pod starts with the volume attached immediately
How to reproduce it:
kubectl apply -f persistent-volumes.yaml --namespace <masked>
Run the helm chart with pod
Actual result:
describe pod msg:
Anything else we need to know?:
Issue is occurs also after we have synced data (rsync to linux vm in azure, with nfs share mounted) into this NFS share.
Once the PV has been mounted and no operations are ongoing the PV and PVC can be removed. 2nd run the pvc is immediately mounted and available. Even if we switch from cluster its fine.
Environment:
kubectl version
): 1.28.3uname -a
):The text was updated successfully, but these errors were encountered: