OCNADD_POD_RESTARTED
Table 5-3 OCNADD_POD_RESTARTED
| Field | Details |
|---|---|
| Triggering Condition | A POD has restarted |
| Severity | Minor |
| Description | A POD has restarted in last 2 min |
| Alert Details |
Summary: 'namespace: {{ "{{" }}$labels.namespace}}, podname: {{ "{{" }}$labels.pod}}, timestamp: {{ "{{" }} with query "time()" }}{{ "{{" }} . | first | value | humanizeTimestamp }}{{ "{{" }} end }}: A Pod has restarted'
Expression: expr: kube_pod_container_status_restarts_total{namespace="{{ .Values.global.cluster.nameSpace.name }}"} > 1 |
| OID | 1.3.6.1.4.1.323.5.3.51.29.5006 |
| Metric Used |
kube_pod_container_status_restarts_total Note: This is a Kubernetes metric used for instance availability monitoring. If the metric is not available, use the similar metric asexposed by the monitoring system. |
| Resolution |
The alert is cleared automatically if the specific pod is up. Steps: 1. Check the application logs. Check for database related failures such as connectivity, Kubernetes secrets, and so on. 2. Run the following command to check orchestration logs for liveness or readiness probe failures: kubectl get po -n <namespace> Note the full name of the pod that is not running, and use it in the following command: kubectl describe pod <desired full pod name> -n <namespace> 3. Check the database status. For more information, see "Oracle Communications Cloud Native Core DBTier User Guide". 4. If the issue persists, capture all the outputs from the above steps and contact unresolvable-reference.html#GUID-6BEC5E06-7EC9-457A-8A82-15663EEAE8AF, If guidance is required. |