-
Notifications
You must be signed in to change notification settings - Fork 2
Open
Description
Recently, we attached a load balancer to a GKE app.
There are several ways that it could fail and it was difficult to track down:
- Is the gateway launched/healthy?
- Is the SSL provisioned properly?
- Is the health check correct?
- Is the port correct?
There is information that sprawls several gcp and k8s to diagnose. Here is a checklist:
- kubectl describe gateway + kubectl describe httproute (Accepted/Programmed/ResolvedRefs)
- kubectl get endpointslice (does the Service actually have endpoints?)
- gcloud compute backend-services get-health … (why unhealthy)
- gcloud compute health-checks describe … (what it’s probing)
- gcloud compute network-endpoint-groups list-network-endpoints … (are endpoints registered?)
Ironically, I only found the solution by navigating to the pod in GCP console.
Some interesting notes:
- If I click on the replica set, there is no useful information
- If I click on the pod revision, there is no useful information -- it navigates to the replica set (wtf?)
- I had to click on the new pod name to find the information under Events tab. (Revealed a readiness probe issue)
Metadata
Metadata
Assignees
Labels
No labels