- Newest
- Most votes
- Most comments
Hi,
You may want to adapt the guidance of this blog post to achieve your goal to export this index: https://aws.amazon.com/blogs/big-data/migrate-your-indexes-to-amazon-opensearch-serverless-with-logstash/
This additional information about serverless logstash will help: https://docs.aws.amazon.com/opensearch-service/latest/developerguide/serverless-clients.html#serverless-logstash
Best,
Didier
FYI for the logstash method mentioned. The logstash-input-opensearch doesn't support opensearch serverless as an input yet. Only as an output. See: https://github.com/opensearch-project/logstash-input-opensearch/issues/53. Did you end up finding a way to do this?
I ended up finding a way to export/import this via a python script. Although not ideal. There's also some api differences I found out between querying serverless and managed that make it harder.
Basically a loop that queries the serverless source using the search_after parameter to fetch data in batches, and bulk indexes it into the destination cluster.
Relevant content
- Accepted Answerasked 9 months ago
- AWS OFFICIALUpdated 4 years ago
- AWS OFFICIALUpdated a year ago
- AWS OFFICIALUpdated 3 years ago
- AWS OFFICIALUpdated 2 years ago
Hi Durand, i mean in openSearch serverless, the inexed data is stored in EBS , right ? Can we export the specific indexed data which stored in EBS into S3 or import specific indexed data which stored in S3 into EBS ? If we can do so, we can use a few OCUs to store many indexed data, since we can export extra data into S3 from EBS if these data is not frequently used.