Cluster Capacity Snapshot
Quantitative node-capacity signals and scheduling headroom captured directly from the Kubernetes bundle.
Operator
Infrastructure
Restore payments-api capacity and stability first: investigate the CrashLoopBackOff pod, fix the failing replica, and get the Deployment back to 3/3 ready before making further changes.
Relieve cluster memory pressure and scheduling blockage: free memory or scale the AKS node pool so pending pods can schedule and the HPA has headroom beyond its current max replicas.
Reduce exposure of the payments namespace: confirm the LoadBalancer is required, then add namespace NetworkPolicies and tighten Service/ServiceAccount settings (automount token and seccomp) for the workload.
Quantitative node-capacity signals and scheduling headroom captured directly from the Kubernetes bundle.
Compact CPU and memory bars so operators can see which nodes are carrying the most pressure without reading raw metrics.
Highest pod-level CPU and memory consumers captured by `kubectl top` during the run.
Autoscaling, disruption, quota, and namespace-default coverage that changes how operators should interpret capacity signals.
Short, operator-oriented callouts for scheduling, rollout, and failing-workload evidence.
A compact operator view of severity and signal distribution before you drop into detailed findings.
Filter the findings table by signal or severity while keeping the current visible count in view.
Detailed review
Expanded explanation for operators who want the model summary after reviewing the findings table.