Error: create: failed to create: Request entity too large: limit is 3145728
Troubleshooting
This section covers common issues that may arise while managing SD Elements and how to address them.
Event 1: Unable to download helm charts
Contact SD Elements support: support@sdelements.com
Event 2: Helm install fails
Refer to current software requirements for helm.
Observe the error message. The following are some common errors and immediate solutions:
Ensure that there are no large files not a part of the helm chart templates found in the path where the helm chart resides.
If this is not sufficient to allow for correction of the issue, contact SD Elements support.
Tip
|
When contacting support, be prepared to share information about your helm, Kubernetes, and SDE helm chart versions. |
Event 3: Helm install has succeeded but one or more pods are showing Statuses such as 'Error' and 'CrashLoopBackoff'
-
Review Events for the pod in question:
$kubectl describe pod {problem_pod_name}
-
Confirm that persistent volume claims are all bound:
$kubectl get pvc NAME STATUS VOLUME CAPACITY ACCESS MODES STORAGECLASS AGE sde1-backup-volume-claim Bound pvc-872c8af0-f258-4ab7-a176-1cf4b07ee588 30Gi RWX microk8s-hostpath 16s sde1-cache-volume-claim Bound pvc-a7f8cf6e-2c8c-44d7-bf35-e8672bdf216e 2Gi RWX microk8s-hostpath 16s sde1-datastore-volume-claim Bound pvc-0d13b63a-9f71-47e0-baba-a0b37af04336 10Gi RWO microk8s-hostpath 16s sde1-db-volume-claim Bound pvc-59126ddc-b697-438b-a6cf-482aba8498f1 30Gi RWO microk8s-hostpath 16s sde1-haystack-volume-claim Bound pvc-b6c02335-b0dc-490f-9027-448ddb8a5367 2Gi RWX microk8s-hostpath 16s sde1-loader-volume-claim Bound pvc-84bd0776-c54f-49db-9e07-7e21581886ac 10Gi RWX microk8s-hostpath 16s sde1-media-volume-claim Bound pvc-7769268e-4faf-4219-86bb-ee2f9b089a26 4Gi RWX microk8s-hostpath 16s sde1-static-volume-claim Bound pvc-af86185c-aa71-4fd5-a254-f30e1845d9d8 4Gi RWX microk8s-hostpath 16s sde1-tls-volume-claim Bound pvc-9ce13a37-0f58-4b2d-8179-a397768de535 10Mi RWX microk8s-hostpath 16s
-
If any have a Status that isn’t 'Bound' or 'Pending', check the Events section for messages that may indicate the cause:
$kubectl describe pvc {problem_pvc_name}
-
-
Review the logs of the problematic pod:
$kubectl logs -f {problem_pod_name}
Event 4: Cron job pod fails with CreateContainerConfigError
-
Review container status for the pod in question:
$ kubectl get pod <PROBLEMATIC_POD> -o jsonpath={.status.containerStatuses}
-
If the state for the container is
waiting
and the message is unable to find a key in the secrets:# Get the corresponding job name for the pod $ kubectl get pod <PROBLEMATIC_POD> -o jsonpath={.metadata.labels.job-name}
# Delete the job corresponding to the problematic pod # kubectl delete job <JOB_NAME>
-