Hi,
I’m currently using K10 in a “no-local, all S3” policy, meaning that every local snapshot made for backups has to be immediatley deleted once it has been exported to S3, as seen here :
I have no problems with backups whatsoever, but I do have a pile (1700+) of VolumeSnapshots that are not deleted.
Every VC looks like :
apiVersion: snapshot.storage.k8s.io/v1
kind: VolumeSnapshotContent
metadata:
annotations:
snapshot.storage.kubernetes.io/volumesnapshot-being-deleted: "yes"
creationTimestamp: "2024-03-09T00:03:40Z"
deletionGracePeriodSeconds: 0
deletionTimestamp: "2024-03-09T00:04:25Z"
finalizers:
- snapshot.storage.kubernetes.io/volumesnapshotcontent-bound-protection
generation: 3
name: snapshot-copy-2b4pthg4-content-f019e565-697e-4649-b797-d461f4c18544
resourceVersion: "175949429"
uid: c0234817-0c1f-4e7d-8969-44fb7a83d5d5
spec:
deletionPolicy: Delete
driver: rook-ceph.rbd.csi.ceph.com
source:
snapshotHandle: 0001-0009-rook-ceph-0000000000000002-8eda267e-e1a2-4242-8836-92df0dbeb325
volumeSnapshotClassName: k10-clone-csi-rbdplugin-snapclass
volumeSnapshotRef:
kind: VolumeSnapshot
name: snapshot-copy-2b4pthg4
namespace: kasten-io
uid: ecfa59d1-7353-4a39-af09-ca304808e9a4
status:
creationTime: 1709942621836570796
readyToUse: true
restoreSize: 0
snapshotHandle: 0001-0009-rook-ceph-0000000000000002-8eda267e-e1a2-4242-8836-92df0dbeb325
Regarding Ceph, the only relevant logs are of course the csi-snapshotter container, which tells :
...
E0419 11:21:58.273585 1 snapshot_controller_base.go:359] could not sync content "snapshot-copy-9jhzn7fz-content-f7e30f08-5d76-4643-9d3f-4604441fb283": failed to delete snapshot "snapshot-copy-9jhzn7fz-content-f7e30f08-5d76-4643-9d3f-4604441fb283", err: failed to delete snapshot content snapshot-copy-9jhzn7fz-content-f7e30f08-5d76-4643-9d3f-4604441fb283: "rpc error: code = InvalidArgument desc = provided secret is empty"
I0419 11:21:58.273685 1 event.go:364] Event(v1.ObjectReference{Kind:"VolumeSnapshotContent", Namespace:"", Name:"snapshot-copy-9jhzn7fz-content-f7e30f08-5d76-4643-9d3f-4604441fb283", UID:"a66cd49d-68ed-4bb3-aa26-efc56e6668ea", APIVersion:"snapshot.storage.k8s.io/v1", ResourceVersion:"204560594", FieldPath:""}): type: 'Warning' reason: 'SnapshotDeleteError' Failed to delete snapshot
...
StorageClass and VolumeSnapshotClass associated with rook-ceph both have secret location annotations.
I do have backups both for RBD and CephFS, and both plugin pods have the same kind of logs
I can provide logs from k10_debug.sh tool if needed.
Any help appreciated, thanks a lot !