S3 performance problems: How can i achieve a s3 request rate above the limit of 3500 PUT per second with multiple prefixes?

1

According to the official document, I try to scale write operations by writing to multiple prefixes, i got nothing, but the same limits(3,500 PUT/COPY/POST/DELETE or 5,500 GET/HEAD requests per second). The official document seems misleading or incorrect.

Q: How can i achieve a s3 request rate above the limit of 3500 PUT per second? Or, how can i make the prefix partitioned ?

PS:I got a solution in the link , but how to "gradually scale up" as described in it ?

已提問 2 年前檢視次數 394 次
1 個回答
0

Throughput on S3 is based on the prefix of the object. The prefix is the initial part of the object name, so for an object with the name upload/data/incoming/file.txt, the prefix is upload/data/incoming/. Each prefix will receive upto the limits you list above, so to increase the 'bucket' throughput, ensure you use unique prefixes.

One thing to remember is that whilst we often think of the components in the prefix as folders, they're not really, so upload is a different prefix to upload/data.

profile picture
已回答 1 年前

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。

回答問題指南