By using AWS re:Post, you agree to the Terms of Use

AWS OpenSearch Service recommended JVM heap size is often set to 50% of system RAM. If JVM heap size is limited to 32GB, does that mean there is no point in using node types with more than 64GB RAM?

0

AWS OpenSearch Service recommended JVM heap size is often set to 50% of system RAM. If JVM heap size is limited to 32GB, does that mean there is no point in using machine types with more than 64GB RAM?

If there is a point in having node types with +64GB (ie. 128GB) why? What is the use of the remaining 32GB for a 64GB RAM/ 96GB for a 128GB RAM after the 32GB has been committed to java heap in a node?

May be we can assume that the operating system and other system tasks uses the remaining memory, but how much of that is really needed by the system tasks or OS?

In a nutshell is there any relevance of having huge memory after 32GB of RAM has been committed to java heap for nodes in an OpenSearch cluster? Please assist if you can. Thanks.

References: https://docs.amazonaws.cn/en_us/opensearch-service/latest/developerguide/auto-tune.html https://aws.amazon.com/blogs/big-data/understanding-the-jvmmemorypressure-metric-changes-in-amazon-opensearch-service/

asked 23 days ago71 views
1 Answers
1
Accepted Answer

in general, after 64GB it is indeed makes more sense to scale out rather that up, however the extra memory is always useful as it will be used for OS cache which in turn will speed up queries if relevant segments are cached by operating system.

answered 23 days ago
  • Thanks for the feedback @Alext_T. That makes sense.

You are not logged in. Log in to post an answer.

A good answer clearly answers the question and provides constructive feedback and encourages professional growth in the question asker.

Guidelines for Answering Questions