-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
400/502/504 errors while doing rollout restart or rolling update #1065
Comments
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
/remove-lifecycle stale |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
/remove-lifecycle stale |
+1 near-exact same situation |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
Stale issues rot after 30d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-testing, kubernetes/test-infra and/or fejta. |
/remove-lifecycle rotten |
Both v1 and v2 controllers supports zero downtime deployment. Need to document how to setup for zero downtime deployment. |
@kishorj You mean |
Issues go stale after 90d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-contributor-experience at kubernetes/community. |
Stale issues rot after 30d of inactivity. If this issue is safe to close now please do so with Send feedback to sig-contributor-experience at kubernetes/community. |
/remove-lifecycle rotten |
you need set up at least preStop sleep 90. |
I tried to fix this PR #1775, but I've ended up with a separate package: https://github.com/foriequal0/pod-graceful-drain |
Filed a documentation issue for setting up zero downtime deployment #2131 |
Hi, I'm getting errors while doing a rolling update and I can reproduce this problem consistently doing a rollout restart. I already tried the recommendations pointed in other issues related with this problem (#814), such as adding a preStop hook to sleep for some seconds so the pods can finish the on going requests but it doesn't solve the problem.
I have also changed the configuration on the load balancer to have the health check interval and threshold-count lower than what I'm setting for the pod's readiness probe so the load balancer could stop sending requests to the pods before receiving the SIGTERM, but without success.
Currently this is the configuration for the ingress, service and deployment:
Ingress
Service
Deployment
To replicate this issue I'm using fortio to call an endpoint on the application constantly during some time (like this:
fortio load -a -c 8 -qps 500 -t 60s "https://app.dev.codacy.org/manual/user/project/dashboard?bid=123"
) and meanwhile I runkubectl rollout restart deployment/codacy-api -n codacy
to restart the pods.In the end there's some errors caused by the rollout restart:
I always get some errors while doing restarts during this test. This is causing some troubles in our application in production when we do rolling updates.
I noticed that in the nginx ingress controller there's the proxy-next-upstream configuration to specify in which cases a request should be passed to the next server, is there any way to do this with this load balancer? Should I use nginx instead?
Thanks for the help.
The text was updated successfully, but these errors were encountered: