Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

OOM-killed core-dump-handler pods #169

Open
yichang-lin opened this issue Jan 15, 2025 · 3 comments
Open

OOM-killed core-dump-handler pods #169

yichang-lin opened this issue Jan 15, 2025 · 3 comments

Comments

@yichang-lin
Copy link

yichang-lin commented Jan 15, 2025

We're using core-dump-handler chart version v9.0.0 and running it on AWS EKS v1.31. Although we have increased request_mem and limit_mem to 256Mi, we are still having core-dump-handler pods get OOM-killed. Is this related to the new EKS version or the new core-dump-handler version? Thanks!

@No9
Copy link
Collaborator

No9 commented Jan 17, 2025

Hi @yichang-lin
Have you looked at the kubelet logs to understand the eviction events in more detail?

@yichang-lin
Copy link
Author

yichang-lin commented Jan 21, 2025

@No9 We haven't checked the logs but the issue was resolved by rolling back the version to v8.10.0.
Here's the detailed log from the OOM-killed pod:

Process core-dump-agent (pid: 954575, oom_score: 132118, oom_score_adj: 998) triggered an OOM kill on itself. The process had reached 65536 pages in size.

@No9
Copy link
Collaborator

No9 commented Jan 21, 2025

Thanks very much - looking at the diff between 9.0.0 and 8.10.0
v8.10.0...v9.0.0
The main changes in the agent was for deployment parameters, a change in the inotify api and a bump in dependencies.
A candidate here is the tokio-scheduler dependency bump.
I'll downgrade the changes and publish 9.1

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants