/var/log/messages may be many GB by itself/var/log/services-logs directory may be more than 10GBVMware Cloud Foundation Automation 9.0.0
Prelude logs that are written to the node are supposed to get truncated if the node disk usage exceeds predefined limit. However, that cleanup code had a bug in 9.0.0 which was fixed in 9.0.1, which prevented it from running to completion in some circumstances.
The issue is resolved in VCFA 9.0.1
Before proceeding, please ensure you have a valid backup or snapshot in case needed logs are being cleared in the below procedure.
Basic things that need to be checked:
Available storage on each node. If it exceeds 75%, delete the content of /var/log/vmsp-logs, /var/log/services-logs and /var/log/messages
rm -rf /var/log/vmsp-logs/* /var/log/services-logs/* truncate -s 1M /var/log/messagesCheck if all nodes are ready. It might take a few minutes for nodes to show up as ready after disk space cleanup. A node replacement might get initiated, check for that.
kubectl get machines -n vmsp-platform
Check if pods in namespaces kube-system and vmsp-platform are running and not in restart loop.
kubectl get pods -n vmsp-platformkubectl get pods -n kube-systemCheck if the object storage and all buckets are not full. This can be done using this command on any node
(need to set KUBECONFIG variable first):
export KUBECONFIG=/etc/kubernetes/admin.confkubectl exec -i -n vmsp-platform seaweedfs-master-0 -- weed shell 2>/dev/null <<< 's3.bucket.list'If utilization of support-bundle bucket is high (>85%), it would need to be cleared. First, generate support bundle to preserve current logs (if haven't done already), then do this:
kubectl exec -it -n vmsp-platform seaweedfs-master-0 -- weed shellfs.rm -r /buckets/support-bundle/fs.cd /buckets/support-bundlefs.rm -r system-logs services-logs events audit apiserver-audit