$kubectl describe pod {problem_pod_name}
Troubleshooting
This section covers common issues that may arise while managing SD Elements and how to address them.
Event 1: Unable to download helm charts
Contact SD Elements support: support@sdelements.com
Event 2: Helm install fails
Observe the error message. If this is not sufficient to allow for correction of the issue, contact SD Elements support.
Event 3: Helm install has succeeded but one or more pods are showing Statuses such as 'Error' and 'CrashLoopBackoff'
-
Review Events for the pod in question:
-
Confirm that persistent volume claims are all bound:
$kubectl get pvc NAME STATUS VOLUME CAPACITY ACCESS MODES STORAGECLASS AGE sde1-backup-volume-claim Bound pvc-872c8af0-f258-4ab7-a176-1cf4b07ee588 30Gi RWX microk8s-hostpath 16s sde1-cache-volume-claim Bound pvc-a7f8cf6e-2c8c-44d7-bf35-e8672bdf216e 2Gi RWX microk8s-hostpath 16s sde1-datastore-volume-claim Bound pvc-0d13b63a-9f71-47e0-baba-a0b37af04336 10Gi RWO microk8s-hostpath 16s sde1-db-volume-claim Bound pvc-59126ddc-b697-438b-a6cf-482aba8498f1 30Gi RWO microk8s-hostpath 16s sde1-haystack-volume-claim Bound pvc-b6c02335-b0dc-490f-9027-448ddb8a5367 2Gi RWX microk8s-hostpath 16s sde1-loader-volume-claim Bound pvc-84bd0776-c54f-49db-9e07-7e21581886ac 10Gi RWX microk8s-hostpath 16s sde1-media-volume-claim Bound pvc-7769268e-4faf-4219-86bb-ee2f9b089a26 4Gi RWX microk8s-hostpath 16s sde1-static-volume-claim Bound pvc-af86185c-aa71-4fd5-a254-f30e1845d9d8 4Gi RWX microk8s-hostpath 16s sde1-tls-volume-claim Bound pvc-9ce13a37-0f58-4b2d-8179-a397768de535 10Mi RWX microk8s-hostpath 16s
-
If any have a Status that isn’t 'Bound' or 'Pending', check the Events section for messages that may indicate the cause:
$kubectl describe pvc {problem_pvc_name}
-
-
Review the logs of the problematic pod:
$kubectl logs -f {problem_pod_name}
Event 4: Cron job pod fails with CreateContainerConfigError
-
Review container status for the pod in question:
$ kubectl get pod <PROBLEMATIC_POD> -o jsonpath={.status.containerStatuses}
-
If the state for the container is
waiting
and the message is unable to find a key in the secrets:# Get the corresponding job name for the pod $ kubectl get pod <PROBLEMATIC_POD> -o jsonpath={.metadata.labels.job-name}
# Delete the job corresponding to the problematic pod # kubectl delete job <JOB_NAME>
-