Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

net-kourier service create timeout while instance rollout with 3 replicas #1278

Open
yaswanthkumar1995 opened this issue Aug 1, 2024 · 2 comments
Labels
lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale.

Comments

@yaswanthkumar1995
Copy link

I have deployment of knative-serving with 3 replicas of knative-serving & 3 replicas of net-kourier contollers with pod anti-affinity such as each pod on each node & pdb of 66% so always 2 pods are available.

probes for knative-serving

    `livenessProbe: failureThreshold: 1 httpGet: path: /health port: probes scheme: HTTP periodSeconds: 5 successThreshold: 1 timeoutSeconds: 1 name: controller ports: - containerPort: 9090 name: metrics protocol: TCP - containerPort: 8008 name: profiling protocol: TCP - containerPort: 8080 name: probes protocol: TCP readinessProbe: failureThreshold: 2 httpGet: path: /readiness port: probes scheme: HTTP periodSeconds: 5 successThreshold: 1 timeoutSeconds: 1`

probes for kourier

    `livenessProbe: failureThreshold: 3 grpc: port: 18000 service: "" periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 name: controller ports: - containerPort: 18000 name: http2-xds protocol: TCP readinessProbe: failureThreshold: 5 grpc: port: 18000 service: "" periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1`

While performing instance rollout with 25% of each time out of 4 instances i am facing timeout issue while creating knative service after 10mins. why doesn't it using other pods which are healthy & running to route the request & not retrying on the backend with healthy pods.

kn service create hello-23 --image=docker.io/johnralston/helloworld --env TARGET="test run" Warning: Kubernetes default value is insecure, Knative may default this to secure in a future release: spec.template.spec.containers[0].securityContext.allowPrivilegeEscalation, spec.template.spec.containers[0].securityContext.capabilities, spec.template.spec.containers[0].securityContext.runAsNonRoot, spec.template.spec.containers[0].securityContext.seccompProfile Creating service 'hello-23' in namespace 'default': **Error: timeout: service 'hello-23' not ready after 600 seconds** Run 'kn --help' for usage

pods
net-kourier-controller-67dc75cd58-blbk8 0/1 ContainerCreating 0 1s <none> ip-node-3.ec2.internal <none> <none> net-kourier-controller-67dc75cd58-hsdzv 1/1 Running 0 58m node-1 ip-node-1.ec2.internal <none> <none> net-kourier-controller-67dc75cd58-tzxn6 1/1 Running 0 4m43s node-2 ip-node-2.ec2.internal <none> <none>

@yaswanthkumar1995
Copy link
Author

net-kourier logs error
{"severity":"ERROR","timestamp":"2024-08-01T16:51:04.670842621Z","logger":"net-kourier-controller","caller":"status/status.go:405","message":"Probing of http://hello-87-default.stage-ee-api.test-domain.com/ failed, IP: 10.208.56.218:8090, ready: false, error: error roundtripping http://hello-87-default.stage-ee-api.test-domain.com/healthz: context deadline exceeded (depth: 0)","commit":"3950a5b-dirty","knative.dev/controller":"knative.dev.net-kourier.pkg.reconciler.ingress.Reconciler","knative.dev/kind":"networking.internal.knative.dev.Ingress","knative.dev/traceid":"46156891-f482-4d8a-8b98-2091d65ae751","knative.dev/key":"default/hello-87","stacktrace":"knative.dev/networking/pkg/status.(*Prober).processWorkItem\n\tknative.dev/[email protected]/pkg/status/status.go:405\nknative.dev/networking/pkg/status.(*Prober).Start.func1\n\tknative.dev/[email protected]/pkg/status/status.go:290"}

Copy link

This issue is stale because it has been open for 90 days with no
activity. It will automatically close after 30 more days of
inactivity. Reopen the issue with /reopen. Mark the issue as
fresh by adding the comment /remove-lifecycle stale.

@github-actions github-actions bot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Oct 31, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale.
Projects
None yet
Development

No branches or pull requests

1 participant