Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Prevent Policy Server Crash in case of a maintenance in Kubernetes Nodepools #383

Closed
wants to merge 4 commits into from

Conversation

ferhatguneri
Copy link
Contributor

@ferhatguneri ferhatguneri commented Feb 7, 2024

Description

If policy server crash because of wrong clusteradmissionpolicy it is blocking pod to be created and cannot evaluate resources correctly. which is affecting control plane. therefore need to keep policy-server always available. These changes also can be added to policy server deployment but since it is hardcoded with Go, I thought of editing the helm chart.

Validate helm chart to have more than 1 replica for policy-server

Signed-off-by: Ferhat Güneri <[email protected]>
increase replica count for policy server

Signed-off-by: Ferhat Güneri <[email protected]>
Signed-off-by: Ferhat Güneri <[email protected]>
@ferhatguneri ferhatguneri requested a review from a team as a code owner February 7, 2024 13:11
@flavio
Copy link
Member

flavio commented Feb 7, 2024

Thanks for the contribution. This fixes only the PolicyServer named default. The proper fix should be done inside of the kubewarden-controller.
This is tracked with kubewarden/kubewarden-controller#564

@flavio
Copy link
Member

flavio commented Feb 7, 2024

Yes, that would be great. Take a look at kubewarden/kubewarden-controller#564 (comment) and implement the "Pod Disruption Budget" section.

Feel free to reach out if something is not clear or if you need help

@ferhatguneri
Copy link
Contributor Author

Hi @flavio I'm aware of that issue but there is no progress since long time. This is a very critical problem and I believe it needs to be fixed immediately. Do you have any idea how long it will take to get it fixed? It is not really good idea to patch these helm charts internally and deal with the upcoming changes.

@flavio
Copy link
Member

flavio commented Feb 7, 2024

I think we can start working on this fix during the next sprint and make it part of the 1.11 release, but I have to discuss that with the other maintainers.

@kkaempf kkaempf added the kind/bug Something isn't working label Feb 8, 2024
@kkaempf kkaempf added this to the v1.11 milestone Feb 8, 2024
Copy link
Member

@viccuad viccuad left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi, many thanks for bringing this forward!

I totally agree with providing a PodDisruptionBudget, yet we have the policy for the default values to not be production ready; people deploy with the defaults to test in a local cluster, and there's a myriad of production deployment flavours that cannot be covered via the default values.

I would welcome an optional configurable setting for the PodDisruptionBudget and the minimum replicaCount.

@flavio
Copy link
Member

flavio commented Feb 9, 2024

Closing, we will fix that inside of the controller with kubewarden/kubewarden-controller#564

@flavio flavio closed this Feb 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
kind/bug Something isn't working
Projects
Archived in project
Development

Successfully merging this pull request may close these issues.

4 participants