Replies: 5 comments
-
When you see the error |
Beta Was this translation helpful? Give feedback.
-
yes, that's true, there is a watcher:
But the rbd-image is not mapped and mounted on this server. Is there anything else possibly creating the watcher?
|
Beta Was this translation helpful? Give feedback.
-
Is it possible that the RBD image is mirrored to an other location? A process like |
Beta Was this translation helpful? Give feedback.
-
I also encountered the same problem, but sometimes it can automatically mount back after a period of time, and sometimes it keeps getting stuck. I want to investigate whether the problem is in the CEPH cluster or in the CSI plugin. |
Beta Was this translation helpful? Give feedback.
-
How to solve it? |
Beta Was this translation helpful? Give feedback.
-
Describe the bug
We have a Proxmox cluster with Ceph and k8s running there.
We are using the csi-rbd provider to deploy PVCs.
It is all seems to be working, a pod is started, pvc is created, a rbd image seems to be created, all is mounted and the app inside a pod is able to use it. Until the pod is deleted and recreated. From that moment, it won't start, as the PVC is not ready for the pod:
I think I was able to verify that the volume is really not mounted anywhere. I got a shell inside each
ceph-csi-rbd-nodeplugin
pod, on each worker node, and from thecsi-rbdplugin
container I was looking for the rbd device with something likerbd device ls | grep csi-vol-dea80a4e-85b6-46e2-9d08-ffba10ef7825
. I double check directly on the worker node OS, by looking at all mount entries. Nothing found.I also tried to mount the device manually. I picked a
ceph-csi-rbd-nodeplugin
container, prepared ceph and keyring config and I was able to map and mount:rbd -c /tmp/ceph.conf -k /tmp/ceph.keyring --id k8s-dev --pool k8s-dev map csi-vol-dea80a4e-85b6-46e2-9d08-ffba10ef7825
, and then mount the block device on the host with like:mount /dev/rbd7 /mnt/
. That worked. Now the rbd image is actually being used :)But I can't find out why k8s things it can't be mounted. I need some help here please.
Is there a change that the
storage.kubernetes.io/csiProvisionerIdentity
could play any role in all of this? Does it change when theceph-csi-rbd-provisioner
is redeployed?Because I noted that the
csiProvisionerIdentity
is different among persistent volumes and I did a couple of redeploys of the provisioner.Environment details
fuse
orkernel
. for rbd itskrbd
orrbd-nbd
) :Steps to reproduce
Init
phase because ofFailedMount
-MountVolume.MountDevice failed for volume "pvc-7e9588ad-19d1-4099-bd49-30fd2b9f0d42" : rpc error: code = Internal desc = rbd image k8s-dev/csi-vol-dea80a4e-85b6-46e2-9d08-ffba10ef7825 is still being used
Nothing interesting from logs unfortunately, or at least I have not noticed.
Beta Was this translation helpful? Give feedback.
All reactions