-
Notifications
You must be signed in to change notification settings - Fork 162
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Memory usage keeps increasing on md5sum compute #502
Comments
Attaching debug logs for reference |
Hi, thank you for the feedback. Consider that Mountpoint is optimized for reading large files sequentially and can prefetch data when it detects a sequential read pattern to improve throughput. This may have an impact on the memory usage, depending on the specific access pattern of different applications. In your use cases with Also, not sure if relevant, but we have an open issue around how dropped GetObject requests (e.g. on out-of-order reads) are handled: #510. It may be worth tracking that and re-run your workflow once it is fixed. |
Thank you for the response. Please check if it is worth considering a command line option or some configuration to limit the maximum prefetch size. This would allow users to set the memory limits for a container more reliably. May be #510 fix will help. Will retry once it is fixed. |
I second that it would be very convenient to have an option to limit the memory usage for scenarios where memory availability is limited. I understand that this will likely impact performance, but that's still better than getting OOM killed. If I wanted to modify that behavior to preload only 128MiB for example, I'd need to modify the constants here right? mountpoint-s3/mountpoint-s3/src/prefetch.rs Line 142 in 7dcaee0
|
@CrawX yes, that's the constant you'd want to modify to scale back the prefetcher's aggressiveness. We're currently looking into a more comprehensive way to limit memory usage; hope to have more to share on that soon! |
I came across this issue while searching for excessive memory usage, my use case is reading 100 large files sequencially and concurrently, I can confirm that after updating |
We don't currently plan to expose this as a configuration. Instead, we're working on improvements that will allow Mountpoint to automatically scale down the prefetching amount based on available resources. I don't have any date I can share for when this would be completed, but the work is ongoing. I hope to be able to share more news soon. (Most recent change refactoring prefetching which prepares for this work: #980) Sorry for the delay in responding here! |
Thanks @dannycjones, this is great news, as a workaround we are using mountpoint by patching the max_request_size value for now. I'll be waiting for your work to be completed. |
I've created this issue which is where we'll share updates on the automatic prefetcher scaling: #987. |
Mountpoint v1.10.0 has been released with some prefetcher improvements and might reduce memory usage. Could you please try upgrading to see if it provides any improvements for you? |
Mountpoint for Amazon S3 version
mount-s3 1.0.1-unofficial+7643a22
AWS Region
us-east-1
Describe the running environment
Running on a local PC docker container. ALso experienced OOMKilled when running as a pod on AWS EKS
What happened?
docker stats
memory usage keeps increasing steadily to 2GB and abovesame behaviour with above python code execution also
Relevant log output
The text was updated successfully, but these errors were encountered: