-
Notifications
You must be signed in to change notification settings - Fork 1.5k
[KEP-5282]Add KEP for Implicit Tolerations #5389
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
cici37
commented
Jun 9, 2025
- One-line PR description:
- Issue link: Implicit tolerations #5282
- Other comments:
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: cici37 The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
/assign @dom4ha @johnbelamaric @sanposhiho |
/wg device-management |
We will take a look, but very likely we cannot make it by the KEP freeze because our bandwidth is limited and there are other on-going KEPs. |
|
||
## Proposal | ||
|
||
Introduce a new PreEnqueue scheduler plugin (or extend the existing TaintToleration plugin) that: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Isn't PreEnqueue too early? In case of PrioritizedAlternatives we may have tainted GPU listed as only one of the options. If we put toleration too early, we (very theoretically) could schedule a pod on tainted node but not use that resource for which the node was tainted.
Alternatively, as mentioned before, tolerations could be also fully implicit (based on policy configuration similar to DeviceTaintRule
) and "applied" invisibly by the TaintToleration plugin during the Filtering phase. However, the problem is that during the Filtering phase in the TaintToleration plugin we may not know which devices got allocated by the DRA plugin, so it may be hard to take decision based on it.
This is why I think that tainting nodes is not the best way to prevent scheduling pods on nodes with scarce resources. Ideally picking or discarding nodes with scarce unused resources should be rather based on scoring mechanisms, or filtering, which could block some placements based on custom policies. So if we forget about taints and toleration mechanism, such logic could be implemented inside DRA plugin.
I agree with @sanposhiho , but also I think there will be ton of code changes in this cycle, which will be very challenging to handle. I'm not convinced that the use case really justifies the effort, as it's not clear to me whether taints and tolerations should be the solution we recommend. |
|
||
Introduce a new PreEnqueue scheduler plugin (or extend the existing TaintToleration plugin) that: | ||
|
||
- Depends on the DRA PreEnqueue plugin to ensure all ResourceClaims and DeviceClasses are resolved. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
what if there are multiple resolutions possible? Will we remove all tolerations so scheduler can decide later where to schedule?
The KEP is discussed in the latest sig scheduling meeting: #5282 (comment) and we'll hold the merge til get an agreement on the use case support. |