-
Notifications
You must be signed in to change notification settings - Fork 4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
VPA Quick OOM not working as expected #7867
Comments
Interesting, I happen to be testing this in a cluster today. I'll leave my findings here |
@adriananeci , please let me know what you find... interestingly we should only get that error message if resourcediff==0
autoscaler/vertical-pod-autoscaler/pkg/updater/priority/update_priority_calculator.go Line 141 in 4e03407
But if we've told it to bump on OOM, then should the resource diff be > 0? |
Turns out the issue I was having is unrelated Going to unassign myself from this for now /unassign |
Hey @ads79 thanks for the detailed logs and parameters! This first line indicates that the updater found a Pod that was OOMKilled recently
Afterwards, it compares the current recommendation against the currently set requests for the Pod and checks if there is a difference. This means: has the So the interesting bit are the logs of the autoscaler/vertical-pod-autoscaler/pkg/recommender/input/oom/observer.go Lines 110 to 115 in ba94506
and then added as new sample to the memory histogram: autoscaler/vertical-pod-autoscaler/pkg/recommender/input/cluster_feeder.go Lines 508 to 512 in ba94506
and autoscaler/vertical-pod-autoscaler/pkg/recommender/model/cluster.go Lines 229 to 243 in ba94506
at container aggregation level you can see your configured min bump values being applied autoscaler/vertical-pod-autoscaler/pkg/recommender/model/container.go Lines 184 to 204 in ba94506
Are you seeing any errors in the |
Which component are you using?: vertical-pod-autoscaler,
/area vertical-pod-autoscaler
What version of the component are you using?:
Component version: 1.3.0 and 1.1.2
What k8s version are you using (
kubectl version
)?:kubectl version
OutputWhat environment is this in?:
AWS EKS
What did you expect to happen?:
Quick OOM detected to bump the values by what is configured:
What happened instead?:
VPA decided to do nothing:
How to reproduce it (as minimally and precisely as possible):
Trace what happens when a pod gets oomkilled.
Anything else we need to know?:
vpa recommender to take into account 1 week of usage.
Initial pod sizes are small 25m CPU and 300MB of RAM, hence the 2x or 512MB bump in ram on OOM
The text was updated successfully, but these errors were encountered: