This KB is designed to help identify which Pod is experiencing the issue.
Once identified, use the steps in the following KB to resolve the issue:
https://knowledge.broadcom.com/external/article?legacyId=96425
Symptoms:
There are 4 main Pods which can experience the issue, each of which have a different method by which you can review the current space usage.
$ kubectl exec alertmanager-tmc-local-monitoring-tmc-local-0 -c alertmanager -- df -h /data
Filesystem Size Used Avail Use% Mounted on
/dev/sdd 2.0G 2.0G 0M 100% /data
Proceed to run KB 96425 replacing any reference to <PACKAGE_INSTALL_NAME> with tmc-local-monitoring.
$ kubectl exec kafka-controller-0 -c kafka -- df -h /bitnami/kafka
Filesystem Size Used Avail Use% Mounted on
/dev/sdc 5G 5G 0M 100% /bitnami/kafka
Proceed to run KB 96425 replacing any reference to <PACKAGE_INSTALL_NAME> with kafka.
$ kubectl exec postgres-postgresql-0 -c postgresql -- df -h /bitnami/postgresql
Filesystem Size Used Avail Use% Mounted on
/dev/sdb 7.8G 7.8G 0M 100% /bitnami/postgresql
Proceed to run KB 96425 replacing any reference to <PACKAGE_INSTALL_NAME> with postgres.
$ kubectl get pods
prometheus-server-tmc-local-monitoring-tmc-local-0 1/2 CrashLoopBackOff 1327 (3m52s ago) 13d
$ kubectl logs prometheus-server-tmc-local-monitoring-tmc-local-0 -c prometheus
ts=2024-01-24T18:51:43.125Z caller=main.go:1166 level=error err="opening storage failed: open /prometheus/wal/00000582: no space left on device"
Proceed to run KB 96425 replacing any reference to <PACKAGE_INSTALL_NAME> with tmc-local-monitoring.