TL;DR apparently the CI automation (flake/edge-case/…) removed three critical files which are responsible for synchronizing the application states across our Kubernetes clusters. Once these are gone, the system reconciles and thinks that all pods should be deleted, which was the case here. The post-mortem will most likely include the reason why that happened and what we’re doing to prevent gitops manifests incorrectly being deleted in the future (most likely locks/improved human review process)